var/home/core/zuul-output/0000755000175000017500000000000015112032035014515 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112045416015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005552144115112045410017675 0ustar rootrootNov 27 11:24:41 crc systemd[1]: Starting Kubernetes Kubelet... Nov 27 11:24:41 crc restorecon[4705]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:41 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 11:24:42 crc restorecon[4705]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 27 11:24:43 crc kubenswrapper[4796]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 11:24:43 crc kubenswrapper[4796]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 27 11:24:43 crc kubenswrapper[4796]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 11:24:43 crc kubenswrapper[4796]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 11:24:43 crc kubenswrapper[4796]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 27 11:24:43 crc kubenswrapper[4796]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.244329 4796 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249448 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249475 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249485 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249495 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249503 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249511 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249519 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249527 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249535 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249543 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249551 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249558 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249566 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249574 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249582 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249590 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249597 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249605 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249615 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249626 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249634 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249642 4796 feature_gate.go:330] unrecognized feature gate: Example Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249650 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249658 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249668 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249677 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249684 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249692 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249699 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249707 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249715 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249723 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249731 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249739 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249747 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249755 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249763 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249770 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249778 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249785 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249793 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249801 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249808 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249816 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249823 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249831 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249838 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249846 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249853 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249861 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249868 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249876 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249883 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249893 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249903 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249912 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249921 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249930 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249938 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249946 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249955 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249963 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249974 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249983 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249991 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.249999 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.250006 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.250014 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.250025 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.250038 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.250048 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250307 4796 flags.go:64] FLAG: --address="0.0.0.0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250327 4796 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250341 4796 flags.go:64] FLAG: --anonymous-auth="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250352 4796 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250363 4796 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250373 4796 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250385 4796 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250395 4796 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250405 4796 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250414 4796 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250424 4796 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250433 4796 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250442 4796 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250451 4796 flags.go:64] FLAG: --cgroup-root="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250460 4796 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250469 4796 flags.go:64] FLAG: --client-ca-file="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250478 4796 flags.go:64] FLAG: --cloud-config="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250487 4796 flags.go:64] FLAG: --cloud-provider="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250496 4796 flags.go:64] FLAG: --cluster-dns="[]" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250508 4796 flags.go:64] FLAG: --cluster-domain="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250517 4796 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250526 4796 flags.go:64] FLAG: --config-dir="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250535 4796 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250544 4796 flags.go:64] FLAG: --container-log-max-files="5" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250555 4796 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250564 4796 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250573 4796 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250583 4796 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250591 4796 flags.go:64] FLAG: --contention-profiling="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250600 4796 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250609 4796 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250618 4796 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250627 4796 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250638 4796 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250646 4796 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250655 4796 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250664 4796 flags.go:64] FLAG: --enable-load-reader="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250673 4796 flags.go:64] FLAG: --enable-server="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250682 4796 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250693 4796 flags.go:64] FLAG: --event-burst="100" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250702 4796 flags.go:64] FLAG: --event-qps="50" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250711 4796 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250720 4796 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250729 4796 flags.go:64] FLAG: --eviction-hard="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250739 4796 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250748 4796 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250757 4796 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250766 4796 flags.go:64] FLAG: --eviction-soft="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250775 4796 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250785 4796 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250793 4796 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250802 4796 flags.go:64] FLAG: --experimental-mounter-path="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250811 4796 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250820 4796 flags.go:64] FLAG: --fail-swap-on="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250829 4796 flags.go:64] FLAG: --feature-gates="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250840 4796 flags.go:64] FLAG: --file-check-frequency="20s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250849 4796 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250860 4796 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250871 4796 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250883 4796 flags.go:64] FLAG: --healthz-port="10248" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250894 4796 flags.go:64] FLAG: --help="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250907 4796 flags.go:64] FLAG: --hostname-override="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250918 4796 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250930 4796 flags.go:64] FLAG: --http-check-frequency="20s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250941 4796 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250952 4796 flags.go:64] FLAG: --image-credential-provider-config="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250986 4796 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.250995 4796 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251005 4796 flags.go:64] FLAG: --image-service-endpoint="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251015 4796 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251024 4796 flags.go:64] FLAG: --kube-api-burst="100" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251034 4796 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251043 4796 flags.go:64] FLAG: --kube-api-qps="50" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251051 4796 flags.go:64] FLAG: --kube-reserved="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251061 4796 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251070 4796 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251079 4796 flags.go:64] FLAG: --kubelet-cgroups="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251087 4796 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251096 4796 flags.go:64] FLAG: --lock-file="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251105 4796 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251114 4796 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251123 4796 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251136 4796 flags.go:64] FLAG: --log-json-split-stream="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251145 4796 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251154 4796 flags.go:64] FLAG: --log-text-split-stream="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251163 4796 flags.go:64] FLAG: --logging-format="text" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251171 4796 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251181 4796 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251189 4796 flags.go:64] FLAG: --manifest-url="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251198 4796 flags.go:64] FLAG: --manifest-url-header="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251210 4796 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251219 4796 flags.go:64] FLAG: --max-open-files="1000000" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251236 4796 flags.go:64] FLAG: --max-pods="110" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251245 4796 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251257 4796 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251298 4796 flags.go:64] FLAG: --memory-manager-policy="None" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251307 4796 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251318 4796 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251327 4796 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251336 4796 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251362 4796 flags.go:64] FLAG: --node-status-max-images="50" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251372 4796 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251381 4796 flags.go:64] FLAG: --oom-score-adj="-999" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251390 4796 flags.go:64] FLAG: --pod-cidr="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251399 4796 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251412 4796 flags.go:64] FLAG: --pod-manifest-path="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251420 4796 flags.go:64] FLAG: --pod-max-pids="-1" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251429 4796 flags.go:64] FLAG: --pods-per-core="0" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251438 4796 flags.go:64] FLAG: --port="10250" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251447 4796 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251456 4796 flags.go:64] FLAG: --provider-id="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251465 4796 flags.go:64] FLAG: --qos-reserved="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251474 4796 flags.go:64] FLAG: --read-only-port="10255" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251483 4796 flags.go:64] FLAG: --register-node="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251492 4796 flags.go:64] FLAG: --register-schedulable="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251501 4796 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251515 4796 flags.go:64] FLAG: --registry-burst="10" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251524 4796 flags.go:64] FLAG: --registry-qps="5" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251533 4796 flags.go:64] FLAG: --reserved-cpus="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251541 4796 flags.go:64] FLAG: --reserved-memory="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251552 4796 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251561 4796 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251570 4796 flags.go:64] FLAG: --rotate-certificates="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251579 4796 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251588 4796 flags.go:64] FLAG: --runonce="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251597 4796 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251606 4796 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251616 4796 flags.go:64] FLAG: --seccomp-default="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251624 4796 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251633 4796 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251642 4796 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251651 4796 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251660 4796 flags.go:64] FLAG: --storage-driver-password="root" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251673 4796 flags.go:64] FLAG: --storage-driver-secure="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251682 4796 flags.go:64] FLAG: --storage-driver-table="stats" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251691 4796 flags.go:64] FLAG: --storage-driver-user="root" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251700 4796 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251709 4796 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251718 4796 flags.go:64] FLAG: --system-cgroups="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251727 4796 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251740 4796 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251749 4796 flags.go:64] FLAG: --tls-cert-file="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251758 4796 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251769 4796 flags.go:64] FLAG: --tls-min-version="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251778 4796 flags.go:64] FLAG: --tls-private-key-file="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251787 4796 flags.go:64] FLAG: --topology-manager-policy="none" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251795 4796 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251804 4796 flags.go:64] FLAG: --topology-manager-scope="container" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251813 4796 flags.go:64] FLAG: --v="2" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251824 4796 flags.go:64] FLAG: --version="false" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251835 4796 flags.go:64] FLAG: --vmodule="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251845 4796 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.251855 4796 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252094 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252105 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252114 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252122 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252133 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252142 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252149 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252157 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252165 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252172 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252180 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252188 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252202 4796 feature_gate.go:330] unrecognized feature gate: Example Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252210 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252218 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252226 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252235 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252243 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252250 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252258 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252293 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252302 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252312 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252320 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252328 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252337 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252345 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252352 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252360 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252367 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252378 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252387 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252398 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252407 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252417 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252425 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252434 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252441 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252449 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252457 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252464 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252473 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252481 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252488 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252498 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252507 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252518 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252528 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252538 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252546 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252555 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252564 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252573 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252582 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252589 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252598 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252607 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252614 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252623 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252632 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252640 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252648 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252656 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252664 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252672 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252679 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252687 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252695 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252702 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252710 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.252717 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.255615 4796 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.268361 4796 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.268426 4796 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268567 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268586 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268597 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268607 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268616 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268626 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268634 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268643 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268652 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268661 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268669 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268677 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268684 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268693 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268701 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268709 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268717 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268725 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268733 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268742 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268750 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268758 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268766 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268773 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268781 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268792 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268803 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268812 4796 feature_gate.go:330] unrecognized feature gate: Example Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268820 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268829 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268838 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268846 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268855 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268865 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268875 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268884 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268893 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268901 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268909 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268917 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268925 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268932 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268940 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268948 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268956 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268965 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268973 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268981 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268988 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.268996 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269003 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269011 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269020 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269029 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269039 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269048 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269057 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269064 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269071 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269079 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269086 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269094 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269103 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269111 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269119 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269130 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269140 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269148 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269156 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269164 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269171 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.269185 4796 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269904 4796 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269924 4796 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269933 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269942 4796 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269951 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269960 4796 feature_gate.go:330] unrecognized feature gate: Example Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269970 4796 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269980 4796 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.269991 4796 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270000 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270008 4796 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270016 4796 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270023 4796 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270034 4796 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270043 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270052 4796 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270060 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270068 4796 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270075 4796 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270083 4796 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270091 4796 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270099 4796 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270107 4796 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270114 4796 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270122 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270130 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270138 4796 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270148 4796 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270158 4796 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270166 4796 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270175 4796 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270183 4796 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270191 4796 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270198 4796 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270206 4796 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270214 4796 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270222 4796 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270229 4796 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270237 4796 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270244 4796 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270251 4796 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270259 4796 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270295 4796 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270305 4796 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270312 4796 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270320 4796 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270328 4796 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270336 4796 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270344 4796 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270352 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270360 4796 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270370 4796 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270379 4796 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270388 4796 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270397 4796 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270405 4796 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270416 4796 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270427 4796 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270437 4796 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270446 4796 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270456 4796 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270466 4796 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270476 4796 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270486 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270497 4796 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270507 4796 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270516 4796 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270526 4796 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270536 4796 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270546 4796 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.270554 4796 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.270567 4796 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.270829 4796 server.go:940] "Client rotation is on, will bootstrap in background" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.285852 4796 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.286007 4796 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.287944 4796 server.go:997] "Starting client certificate rotation" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.288001 4796 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.288250 4796 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-11 06:45:01.376345613 +0000 UTC Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.288381 4796 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 331h20m18.08797047s for next certificate rotation Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.326046 4796 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.328690 4796 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.370362 4796 log.go:25] "Validated CRI v1 runtime API" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.422206 4796 log.go:25] "Validated CRI v1 image API" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.424688 4796 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.433037 4796 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-27-11-20-03-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.433092 4796 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.459357 4796 manager.go:217] Machine: {Timestamp:2025-11-27 11:24:43.457079137 +0000 UTC m=+0.975398125 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:71d23547-b1b3-4287-9b6a-deece17c4b3f BootID:3e2096ad-1d72-4cc4-93b3-695942a724e4 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:2e:c9:e9 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:2e:c9:e9 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:22:6f:d6 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:33:72:79 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c4:0d:bc Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:da:c2:02 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e2:cc:2e:35:fb:b7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:12:ad:23:69:ab:4b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.459796 4796 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.460082 4796 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.462828 4796 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.463119 4796 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.463170 4796 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.463480 4796 topology_manager.go:138] "Creating topology manager with none policy" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.463497 4796 container_manager_linux.go:303] "Creating device plugin manager" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.464355 4796 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.465687 4796 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.465935 4796 state_mem.go:36] "Initialized new in-memory state store" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.466557 4796 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.474235 4796 kubelet.go:418] "Attempting to sync node with API server" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.474346 4796 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.474471 4796 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.474501 4796 kubelet.go:324] "Adding apiserver pod source" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.474550 4796 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.487726 4796 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.489485 4796 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.490124 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.490145 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.490310 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.490315 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.493216 4796 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495582 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495615 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495624 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495633 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495647 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495658 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495669 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495701 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495712 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495721 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495733 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.495744 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.498055 4796 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.498452 4796 server.go:1280] "Started kubelet" Nov 27 11:24:43 crc systemd[1]: Started Kubernetes Kubelet. Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.501437 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.502023 4796 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.502062 4796 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.502885 4796 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504244 4796 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504359 4796 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504411 4796 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 19:18:35.08686853 +0000 UTC Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504466 4796 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 823h53m51.582404834s for next certificate rotation Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504469 4796 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504498 4796 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.504522 4796 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.504768 4796 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.505196 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.505360 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.505712 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="200ms" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.505871 4796 factory.go:55] Registering systemd factory Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.505901 4796 factory.go:221] Registration of the systemd container factory successfully Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.506214 4796 factory.go:153] Registering CRI-O factory Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.506234 4796 factory.go:221] Registration of the crio container factory successfully Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.506315 4796 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.506342 4796 factory.go:103] Registering Raw factory Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.506359 4796 manager.go:1196] Started watching for new ooms in manager Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.506961 4796 manager.go:319] Starting recovery of all containers Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.509213 4796 server.go:460] "Adding debug handlers to kubelet server" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.515576 4796 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.145:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187bd95f8f817c42 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 11:24:43.498429506 +0000 UTC m=+1.016748434,LastTimestamp:2025-11-27 11:24:43.498429506 +0000 UTC m=+1.016748434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519798 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519883 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519929 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519940 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519951 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519963 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519974 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.519987 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520000 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520010 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520021 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520032 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520043 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520055 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520065 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520074 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520083 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520096 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520106 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520114 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520126 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520136 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520146 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520159 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520169 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520179 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520191 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520201 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520211 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520220 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520232 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520242 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520251 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520275 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520313 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520324 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520333 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520343 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520353 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520382 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520393 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520404 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.520415 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521005 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521019 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521064 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521079 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521642 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521689 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521702 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521711 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521721 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521733 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521744 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521754 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521765 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521775 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521783 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521794 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521804 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521814 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521827 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521837 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521847 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521858 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521869 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521888 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521903 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521916 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521926 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521936 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521946 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521956 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521965 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521974 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521986 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.521996 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522007 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522016 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522028 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522038 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522048 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522057 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522066 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522076 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522086 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522096 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522107 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522116 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522126 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522136 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522144 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522153 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522164 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522172 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522181 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522189 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522198 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522207 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522218 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522227 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522237 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522246 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522257 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522294 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522308 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522319 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522330 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522340 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522353 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522366 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522379 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522388 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522399 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522408 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522417 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522426 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522435 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522444 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522455 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522465 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522475 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522485 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522494 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522505 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522518 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522530 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522540 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522552 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522561 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522572 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522582 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522613 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522625 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522634 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522644 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522655 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522665 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522674 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522684 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522695 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522706 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522715 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522724 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.522734 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.524925 4796 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.524996 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525017 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525033 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525051 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525065 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525080 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525094 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525108 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525181 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525194 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525208 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525226 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525238 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525250 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525282 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525298 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525315 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525331 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525346 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525361 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525376 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525390 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525404 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525418 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525432 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525445 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525461 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525480 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525492 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525505 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525518 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525533 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525547 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525559 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525572 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525586 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525601 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525614 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525629 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525643 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525656 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525670 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525687 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525701 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525735 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525749 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525763 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525776 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525789 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525803 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525819 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525836 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525849 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525868 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525881 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525894 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525907 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525921 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525935 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525950 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525963 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525979 4796 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.525992 4796 reconstruct.go:97] "Volume reconstruction finished" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.526005 4796 reconciler.go:26] "Reconciler: start to sync state" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.534595 4796 manager.go:324] Recovery completed Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.546235 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.548110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.548151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.548166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.549015 4796 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.549063 4796 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.549085 4796 state_mem.go:36] "Initialized new in-memory state store" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.566279 4796 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.567806 4796 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.567845 4796 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.567873 4796 kubelet.go:2335] "Starting kubelet main sync loop" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.567914 4796 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 27 11:24:43 crc kubenswrapper[4796]: W1127 11:24:43.569032 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.569099 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.588777 4796 policy_none.go:49] "None policy: Start" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.589578 4796 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.589699 4796 state_mem.go:35] "Initializing new in-memory state store" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.604939 4796 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.657392 4796 manager.go:334] "Starting Device Plugin manager" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.657442 4796 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.657454 4796 server.go:79] "Starting device plugin registration server" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.657858 4796 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.657909 4796 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.658077 4796 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.658155 4796 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.658161 4796 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.663969 4796 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.668557 4796 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.668643 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.669626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.669666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.669679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.669784 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670072 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670129 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670413 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670549 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670641 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.670670 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.673309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.673371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674073 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.673380 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674186 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674242 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674330 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.674389 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675247 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675282 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675333 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675366 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675562 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.675588 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678446 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678471 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678479 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678698 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678709 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.678741 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.681253 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.681314 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.681329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.706639 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="400ms" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.728723 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.728812 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.728873 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.728933 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.728997 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729027 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729111 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729171 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729224 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729248 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729331 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729378 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729415 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729494 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.729565 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.758585 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.759928 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.759982 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.760000 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.760035 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.760670 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.145:6443: connect: connection refused" node="crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831307 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831527 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831555 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831624 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831640 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831758 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831813 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831832 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831879 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831890 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.831957 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832000 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832081 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832087 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832132 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832195 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832156 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832302 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832355 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832355 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832402 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832404 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832437 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832461 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832474 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832521 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.832600 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.961458 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.962686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.962721 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.962732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:43 crc kubenswrapper[4796]: I1127 11:24:43.962751 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:24:43 crc kubenswrapper[4796]: E1127 11:24:43.963161 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.145:6443: connect: connection refused" node="crc" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.012208 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.017277 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.044968 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.071488 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.077887 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 11:24:44 crc kubenswrapper[4796]: E1127 11:24:44.108302 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="800ms" Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.169742 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ac607c02a483540299fc1b19f966038376f5106a0e6799f200e532d7b39e0d5d WatchSource:0}: Error finding container ac607c02a483540299fc1b19f966038376f5106a0e6799f200e532d7b39e0d5d: Status 404 returned error can't find the container with id ac607c02a483540299fc1b19f966038376f5106a0e6799f200e532d7b39e0d5d Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.177953 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2dd14203e04d01fe0081886c0e5fbb74145c1e60c918a411bb0ca7915d875a79 WatchSource:0}: Error finding container 2dd14203e04d01fe0081886c0e5fbb74145c1e60c918a411bb0ca7915d875a79: Status 404 returned error can't find the container with id 2dd14203e04d01fe0081886c0e5fbb74145c1e60c918a411bb0ca7915d875a79 Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.179824 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-22ff398326091b50c6a5a199dae33e7bdc19e51a3876c342584a99296240724f WatchSource:0}: Error finding container 22ff398326091b50c6a5a199dae33e7bdc19e51a3876c342584a99296240724f: Status 404 returned error can't find the container with id 22ff398326091b50c6a5a199dae33e7bdc19e51a3876c342584a99296240724f Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.185253 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e6fb99a6d7d1861dc0c0511a0fdfedd12717f103bd342778fac9c1edcbdbe458 WatchSource:0}: Error finding container e6fb99a6d7d1861dc0c0511a0fdfedd12717f103bd342778fac9c1edcbdbe458: Status 404 returned error can't find the container with id e6fb99a6d7d1861dc0c0511a0fdfedd12717f103bd342778fac9c1edcbdbe458 Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.186560 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e50d303b3064c43d34a5e3b12a6a2517628e777303785ebacef49e156b589ce3 WatchSource:0}: Error finding container e50d303b3064c43d34a5e3b12a6a2517628e777303785ebacef49e156b589ce3: Status 404 returned error can't find the container with id e50d303b3064c43d34a5e3b12a6a2517628e777303785ebacef49e156b589ce3 Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.363705 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.365208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.365293 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.365309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.365395 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:24:44 crc kubenswrapper[4796]: E1127 11:24:44.365964 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.145:6443: connect: connection refused" node="crc" Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.418911 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:44 crc kubenswrapper[4796]: E1127 11:24:44.419011 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.436000 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:44 crc kubenswrapper[4796]: E1127 11:24:44.436096 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:44 crc kubenswrapper[4796]: W1127 11:24:44.466319 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:44 crc kubenswrapper[4796]: E1127 11:24:44.466491 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.503676 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.573312 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e50d303b3064c43d34a5e3b12a6a2517628e777303785ebacef49e156b589ce3"} Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.574589 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e6fb99a6d7d1861dc0c0511a0fdfedd12717f103bd342778fac9c1edcbdbe458"} Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.575716 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"22ff398326091b50c6a5a199dae33e7bdc19e51a3876c342584a99296240724f"} Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.576718 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2dd14203e04d01fe0081886c0e5fbb74145c1e60c918a411bb0ca7915d875a79"} Nov 27 11:24:44 crc kubenswrapper[4796]: I1127 11:24:44.577576 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ac607c02a483540299fc1b19f966038376f5106a0e6799f200e532d7b39e0d5d"} Nov 27 11:24:44 crc kubenswrapper[4796]: E1127 11:24:44.909661 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="1.6s" Nov 27 11:24:45 crc kubenswrapper[4796]: W1127 11:24:45.013909 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:45 crc kubenswrapper[4796]: E1127 11:24:45.014028 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.166465 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.169073 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.169134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.169148 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.169183 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:24:45 crc kubenswrapper[4796]: E1127 11:24:45.170392 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.145:6443: connect: connection refused" node="crc" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.503505 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.587435 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7"} Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.587513 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83"} Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.590523 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964" exitCode=0 Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.590605 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964"} Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.590681 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.592045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.592080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.592093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.592965 4796 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac" exitCode=0 Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.593036 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.593038 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac"} Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.594054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.594110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.594134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.594519 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.595316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.595342 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.595353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.596240 4796 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e" exitCode=0 Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.596380 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e"} Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.596460 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.597557 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.597623 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.597701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.600554 4796 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a" exitCode=0 Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.600594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a"} Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.600670 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.602104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.602163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:45 crc kubenswrapper[4796]: I1127 11:24:45.602185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:46 crc kubenswrapper[4796]: W1127 11:24:46.053576 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:46 crc kubenswrapper[4796]: E1127 11:24:46.053646 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:46 crc kubenswrapper[4796]: W1127 11:24:46.124939 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:46 crc kubenswrapper[4796]: E1127 11:24:46.125023 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.502442 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:46 crc kubenswrapper[4796]: E1127 11:24:46.510887 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="3.2s" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.604487 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.604617 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.605578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.605614 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.605626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.608585 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.608635 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.608653 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.608602 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.609583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.609624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.609657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.613500 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.613535 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.613618 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.614969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.615029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.615044 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.617741 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.617834 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.617848 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.617862 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.620812 4796 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20" exitCode=0 Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.620870 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20"} Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.620982 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.622407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.622442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.622457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:46 crc kubenswrapper[4796]: W1127 11:24:46.697920 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:46 crc kubenswrapper[4796]: E1127 11:24:46.698016 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.771146 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.772966 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.773033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.773049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:46 crc kubenswrapper[4796]: I1127 11:24:46.773086 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:24:46 crc kubenswrapper[4796]: E1127 11:24:46.776949 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.145:6443: connect: connection refused" node="crc" Nov 27 11:24:47 crc kubenswrapper[4796]: W1127 11:24:47.094211 4796 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.145:6443: connect: connection refused Nov 27 11:24:47 crc kubenswrapper[4796]: E1127 11:24:47.094325 4796 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.145:6443: connect: connection refused" logger="UnhandledError" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.540519 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.607825 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.617114 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.629138 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1981a20a04c881ba18485110465aeda59942a35f5abbfe29709eb20c8afe9a53"} Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.629410 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.630988 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.631037 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.631050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.632877 4796 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17" exitCode=0 Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.632991 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.633000 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.633014 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.633048 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17"} Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.633024 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.633131 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634552 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634571 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634824 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:47 crc kubenswrapper[4796]: I1127 11:24:47.634865 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.640809 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484"} Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.640856 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b"} Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.640866 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18"} Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.640876 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee"} Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.640855 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.640918 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.641037 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.641147 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.641646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.641679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.641693 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.642389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.642417 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.642427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.664027 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.664332 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.665435 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.665472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.665482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:48 crc kubenswrapper[4796]: I1127 11:24:48.834371 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.649966 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279"} Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.650089 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.650142 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.650157 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.650777 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.651453 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.651483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.651494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.651651 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.651730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.651745 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.652731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.652777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.652792 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.977505 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.979092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.979129 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.979143 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:49 crc kubenswrapper[4796]: I1127 11:24:49.979167 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.225075 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.541400 4796 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.541493 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.565723 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.653082 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.653193 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.653240 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654688 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654743 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654817 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.654850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:50 crc kubenswrapper[4796]: I1127 11:24:50.666876 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 27 11:24:51 crc kubenswrapper[4796]: I1127 11:24:51.655934 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:51 crc kubenswrapper[4796]: I1127 11:24:51.657076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:51 crc kubenswrapper[4796]: I1127 11:24:51.657141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:51 crc kubenswrapper[4796]: I1127 11:24:51.657170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:52 crc kubenswrapper[4796]: I1127 11:24:52.493697 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:24:52 crc kubenswrapper[4796]: I1127 11:24:52.493982 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:52 crc kubenswrapper[4796]: I1127 11:24:52.495522 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:52 crc kubenswrapper[4796]: I1127 11:24:52.495559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:52 crc kubenswrapper[4796]: I1127 11:24:52.495573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:53 crc kubenswrapper[4796]: E1127 11:24:53.664570 4796 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 11:24:56 crc kubenswrapper[4796]: I1127 11:24:56.376842 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 27 11:24:56 crc kubenswrapper[4796]: I1127 11:24:56.377165 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:56 crc kubenswrapper[4796]: I1127 11:24:56.378730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:56 crc kubenswrapper[4796]: I1127 11:24:56.378794 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:56 crc kubenswrapper[4796]: I1127 11:24:56.378808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.407726 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:54842->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.407795 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:54842->192.168.126.11:17697: read: connection reset by peer" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.502753 4796 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.676998 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.678599 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1981a20a04c881ba18485110465aeda59942a35f5abbfe29709eb20c8afe9a53" exitCode=255 Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.678635 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1981a20a04c881ba18485110465aeda59942a35f5abbfe29709eb20c8afe9a53"} Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.678773 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.679853 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.679883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.679891 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:24:57 crc kubenswrapper[4796]: I1127 11:24:57.680407 4796 scope.go:117] "RemoveContainer" containerID="1981a20a04c881ba18485110465aeda59942a35f5abbfe29709eb20c8afe9a53" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.394656 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.394763 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.399261 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.399355 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.683466 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.684991 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1"} Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.685132 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.685854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.685878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:24:58 crc kubenswrapper[4796]: I1127 11:24:58.685886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.233388 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.233635 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.233784 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.235092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.235137 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.235146 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.238644 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.541605 4796 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded" start-of-body= Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.541765 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.574401 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.574677 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.576483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.576525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.576541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.690870 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.692501 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.692595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:00 crc kubenswrapper[4796]: I1127 11:25:00.692620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:01 crc kubenswrapper[4796]: I1127 11:25:01.693492 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:25:01 crc kubenswrapper[4796]: I1127 11:25:01.694791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:01 crc kubenswrapper[4796]: I1127 11:25:01.694861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:01 crc kubenswrapper[4796]: I1127 11:25:01.694888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.392546 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.395337 4796 trace.go:236] Trace[910260437]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 11:24:49.313) (total time: 14081ms): Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[910260437]: ---"Objects listed" error: 14081ms (11:25:03.395) Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[910260437]: [14.08126412s] [14.08126412s] END Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.395376 4796 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.397994 4796 trace.go:236] Trace[1174012835]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 11:24:52.806) (total time: 10591ms): Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[1174012835]: ---"Objects listed" error: 10591ms (11:25:03.397) Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[1174012835]: [10.591134852s] [10.591134852s] END Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.398052 4796 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.398033 4796 trace.go:236] Trace[211608181]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 11:24:50.956) (total time: 12441ms): Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[211608181]: ---"Objects listed" error: 12441ms (11:25:03.397) Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[211608181]: [12.441097057s] [12.441097057s] END Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.398085 4796 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.398110 4796 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.399478 4796 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.402327 4796 trace.go:236] Trace[590089723]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 11:24:52.571) (total time: 10830ms): Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[590089723]: ---"Objects listed" error: 10830ms (11:25:03.402) Nov 27 11:25:03 crc kubenswrapper[4796]: Trace[590089723]: [10.830380801s] [10.830380801s] END Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.402360 4796 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.484868 4796 apiserver.go:52] "Watching apiserver" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.499440 4796 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.499989 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.500440 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.500643 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.500775 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.500830 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.500923 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.501023 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.501186 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.501249 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.501319 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503459 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503592 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503609 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503667 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503732 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503785 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503826 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.503972 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.504130 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.506527 4796 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.561423 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.580746 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600621 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600675 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600705 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600727 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600753 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600775 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600795 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600852 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600876 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600898 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600918 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600940 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600962 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.600983 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601029 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601028 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601074 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601097 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601141 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601166 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601190 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601182 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601212 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601233 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601256 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601323 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601346 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601376 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601399 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601452 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601477 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601501 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601555 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601580 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601603 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601629 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601652 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601683 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601711 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601733 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601755 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601777 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601815 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601837 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601859 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601884 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601909 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601932 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601956 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601979 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602004 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602029 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602051 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602074 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602116 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602140 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602165 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602188 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602210 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602232 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602254 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602326 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602348 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602369 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602391 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602411 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602433 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602457 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602480 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602502 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602523 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602543 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602569 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602591 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602614 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602636 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602657 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602713 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602736 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602759 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602781 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602801 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602824 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602852 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602873 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602895 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602919 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602942 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602963 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602987 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603010 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603033 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603056 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603079 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603103 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603125 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603148 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603172 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603195 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603217 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603242 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603280 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603303 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603326 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603348 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603371 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603393 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603417 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603438 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601321 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603461 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601435 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601455 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601604 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601620 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601660 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601678 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.601869 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602121 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602134 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602165 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602179 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602239 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602419 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603574 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602440 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602461 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602548 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603601 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602633 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603607 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602679 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602823 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602866 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602916 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.602921 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603093 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603090 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603113 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603110 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603448 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603623 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603746 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603767 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603789 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603809 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603828 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603844 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603841 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603870 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603861 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603947 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603964 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.603995 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604028 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604110 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604122 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604141 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604148 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604163 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604168 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604205 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604235 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604293 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604303 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604330 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604336 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604378 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604340 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604382 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604381 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604440 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604462 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604480 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604503 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604537 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604562 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604579 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604596 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604616 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604638 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604647 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604654 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604698 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604726 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604735 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604750 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604812 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604836 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604860 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604883 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604909 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604932 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604956 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604981 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605004 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605027 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605051 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605074 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605096 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605122 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605145 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605168 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605191 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605213 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605237 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605260 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605699 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605724 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605748 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605771 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605793 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605818 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605840 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605862 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605884 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605906 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605928 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605950 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605974 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605997 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606019 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606041 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606065 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606088 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606111 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606135 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606175 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606198 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606221 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606244 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606285 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606308 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606331 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606352 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606374 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606397 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606419 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606441 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606464 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606486 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606508 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606556 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606588 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606615 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606640 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606663 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606690 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606713 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606790 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606819 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606846 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606873 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606906 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606968 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606983 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.606997 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607011 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607026 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607039 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607053 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607067 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607081 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607093 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607107 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607121 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607136 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607149 4796 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607162 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607176 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607190 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607208 4796 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607221 4796 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607234 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607247 4796 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607260 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607291 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607306 4796 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607320 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607334 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607348 4796 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607362 4796 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607376 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607390 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607404 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607418 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607431 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607444 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607457 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607472 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607486 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607500 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607512 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607526 4796 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607540 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607553 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607567 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607581 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607599 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.607612 4796 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.628303 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.650493 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.653090 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604772 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604796 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604815 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604933 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.604993 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605046 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605071 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605089 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605204 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605281 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.605317 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.620184 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.620251 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.620401 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.620620 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.620962 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.621128 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.622248 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.622636 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.622767 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.622921 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.623120 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.623507 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.623763 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.623800 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.623946 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.623973 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.624382 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.624383 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.624857 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.625566 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.626060 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.626377 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.626745 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.627381 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.627498 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.627683 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.627768 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.627791 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.628711 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.628961 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.629464 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.629596 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.629879 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.629900 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.630691 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.631479 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.631694 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.631894 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.632085 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.633575 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.642925 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.643282 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.643471 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.647507 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.647712 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.647751 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.648139 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.648350 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.648839 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.649176 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.649373 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.649556 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.649840 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.650000 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.650259 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.650292 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.650735 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.653226 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.653453 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.653478 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.654816 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:04.15478854 +0000 UTC m=+21.673107458 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.653750 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.654863 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:04.154851602 +0000 UTC m=+21.673170520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.654076 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.654050 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.653966 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.661796 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.662252 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.663313 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.663723 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.664085 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.664103 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.664309 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.666126 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.666829 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.667549 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.667674 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.668178 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.668258 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.668458 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.668512 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.670796 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.670842 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.672910 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.675716 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.676412 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.677407 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.677526 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.677539 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.677758 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.678392 4796 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.678429 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.678674 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.678700 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.679717 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.679793 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.680036 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.680401 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.680841 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.680906 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.681420 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.682074 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.682084 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.682513 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.682522 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.682776 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683350 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683493 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683651 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683742 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683829 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683968 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.683243 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.684147 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.684475 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:25:04.184453803 +0000 UTC m=+21.702772711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.684922 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.686790 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.690491 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.691280 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.691711 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.692972 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.693462 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.693937 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.694375 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.694689 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.695085 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.695477 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.696022 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.696137 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.696197 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.696787 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.699051 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.699084 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.699104 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.699176 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:04.19915285 +0000 UTC m=+21.717471978 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.699318 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.702853 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.703491 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.704416 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.706508 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.707236 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708415 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708534 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708590 4796 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708607 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708621 4796 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708633 4796 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708645 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708658 4796 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708668 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708680 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708691 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708701 4796 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708711 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708722 4796 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708734 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708745 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708758 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708772 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708784 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708795 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708806 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708816 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708828 4796 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708839 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708850 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708860 4796 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708873 4796 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708885 4796 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708897 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708908 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708919 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708929 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708940 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708949 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708959 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708970 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708980 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.708991 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709000 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709011 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709022 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709032 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709044 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709064 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709075 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709085 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709095 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709105 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709116 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709126 4796 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709136 4796 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709147 4796 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709157 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709169 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709179 4796 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709189 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709204 4796 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709214 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709223 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709234 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709245 4796 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709255 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709283 4796 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709309 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709320 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709329 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709340 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709350 4796 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709369 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709379 4796 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709389 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709404 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709416 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709428 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709438 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709451 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709463 4796 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709473 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709482 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709493 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709503 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709514 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709525 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709535 4796 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709545 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709556 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709566 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709577 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709587 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709597 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709607 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709618 4796 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709628 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709639 4796 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709663 4796 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709673 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709685 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709697 4796 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709707 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709717 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709728 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709750 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709762 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709772 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709782 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709794 4796 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709806 4796 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709817 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709828 4796 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709838 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709850 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709860 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709870 4796 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709879 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709890 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709899 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709910 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709920 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709930 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709939 4796 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709949 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709959 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709971 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.709994 4796 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710004 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710014 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710024 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710034 4796 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710044 4796 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710055 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710065 4796 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710076 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710087 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710100 4796 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710111 4796 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710121 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710131 4796 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710142 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710153 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710163 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710173 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710183 4796 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710195 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710206 4796 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710217 4796 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710227 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710333 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.710360 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.712141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.714106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.716311 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.717164 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.717468 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.719640 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.719856 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.724985 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.725049 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.725066 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:03 crc kubenswrapper[4796]: E1127 11:25:03.725148 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:04.225124492 +0000 UTC m=+21.743443410 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.727196 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.752784 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.753172 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.763764 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.775118 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.785160 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.794016 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.803570 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.810581 4796 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.810625 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.810642 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.810657 4796 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.815743 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.817841 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 11:25:03 crc kubenswrapper[4796]: W1127 11:25:03.829479 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-247c25737b551e5d401c6ec9040177dbbf01a423ecd99aa50c10c6c5d8c3c800 WatchSource:0}: Error finding container 247c25737b551e5d401c6ec9040177dbbf01a423ecd99aa50c10c6c5d8c3c800: Status 404 returned error can't find the container with id 247c25737b551e5d401c6ec9040177dbbf01a423ecd99aa50c10c6c5d8c3c800 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.831024 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.834019 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.840118 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 11:25:03 crc kubenswrapper[4796]: W1127 11:25:03.843048 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-8f827125649ca083488bb54f90a30388f74b5be18f8cb2a32020be93724250e2 WatchSource:0}: Error finding container 8f827125649ca083488bb54f90a30388f74b5be18f8cb2a32020be93724250e2: Status 404 returned error can't find the container with id 8f827125649ca083488bb54f90a30388f74b5be18f8cb2a32020be93724250e2 Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.848964 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:03 crc kubenswrapper[4796]: W1127 11:25:03.853125 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-39f563a9dd65b35062dd0f856190eb218b39ce72c42f9db507cdae5f7828622a WatchSource:0}: Error finding container 39f563a9dd65b35062dd0f856190eb218b39ce72c42f9db507cdae5f7828622a: Status 404 returned error can't find the container with id 39f563a9dd65b35062dd0f856190eb218b39ce72c42f9db507cdae5f7828622a Nov 27 11:25:03 crc kubenswrapper[4796]: I1127 11:25:03.861295 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.213993 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.214054 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.214085 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.214108 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214217 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214232 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214242 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214298 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:05.21428478 +0000 UTC m=+22.732603698 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214617 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:25:05.214608529 +0000 UTC m=+22.732927447 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214669 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214689 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:05.214683851 +0000 UTC m=+22.733002759 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214715 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.214732 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:05.214728002 +0000 UTC m=+22.733046920 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.315137 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.315295 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.315320 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.315333 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.315388 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:05.315372254 +0000 UTC m=+22.833691172 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.568358 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.568476 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.708941 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.709428 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.711067 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" exitCode=255 Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.711125 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.711188 4796 scope.go:117] "RemoveContainer" containerID="1981a20a04c881ba18485110465aeda59942a35f5abbfe29709eb20c8afe9a53" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.712260 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"39f563a9dd65b35062dd0f856190eb218b39ce72c42f9db507cdae5f7828622a"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.713850 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.713884 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.713899 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8f827125649ca083488bb54f90a30388f74b5be18f8cb2a32020be93724250e2"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.715632 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.715663 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"247c25737b551e5d401c6ec9040177dbbf01a423ecd99aa50c10c6c5d8c3c800"} Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.724844 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.738449 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.750302 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.760920 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.775217 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.792773 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.809351 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.832611 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.832694 4796 scope.go:117] "RemoveContainer" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:25:04 crc kubenswrapper[4796]: E1127 11:25:04.832870 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.841188 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.912952 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.944189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.965939 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:04 crc kubenswrapper[4796]: I1127 11:25:04.981798 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:04Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.222405 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.222586 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:25:07.222549005 +0000 UTC m=+24.740867923 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.222672 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.222711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.222752 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.222859 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.222919 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:07.222905525 +0000 UTC m=+24.741224533 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.223006 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.223043 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.223068 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.223178 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:07.223155561 +0000 UTC m=+24.741474519 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.223232 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.223311 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:07.223293445 +0000 UTC m=+24.741612403 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.320664 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.323935 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.324124 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.324144 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.324156 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.324220 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:07.324206344 +0000 UTC m=+24.842525262 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.568802 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.568927 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.569311 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.569368 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.572455 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.573119 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.574071 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.574817 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.575561 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.576959 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.577596 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.578564 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.579167 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.580205 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.580710 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.581773 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.582256 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.582858 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.583760 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.584320 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.585218 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.585704 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.586316 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.589137 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.589782 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.590738 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.591153 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.592243 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.592712 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.593348 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.594620 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.595202 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.596171 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.596621 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.597449 4796 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.597545 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.599249 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.600521 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.601042 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.602906 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.603821 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.604830 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.605589 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.606644 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.607078 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.608061 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.608759 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.609698 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.610187 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.611037 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.611547 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.612619 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.613100 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.613928 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.614395 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.615330 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.615923 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.616436 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.673376 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-ns7ft"] Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.673700 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.675469 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.679637 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.679543 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.687826 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.703669 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.716189 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.720008 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.722078 4796 scope.go:117] "RemoveContainer" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:25:05 crc kubenswrapper[4796]: E1127 11:25:05.722236 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.726970 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.740465 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.758115 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1981a20a04c881ba18485110465aeda59942a35f5abbfe29709eb20c8afe9a53\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:24:57Z\\\",\\\"message\\\":\\\"W1127 11:24:46.850300 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 11:24:46.850826 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764242686 cert, and key in /tmp/serving-cert-1282519371/serving-signer.crt, /tmp/serving-cert-1282519371/serving-signer.key\\\\nI1127 11:24:47.228348 1 observer_polling.go:159] Starting file observer\\\\nW1127 11:24:47.232360 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 11:24:47.232537 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:24:47.233309 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1282519371/tls.crt::/tmp/serving-cert-1282519371/tls.key\\\\\\\"\\\\nF1127 11:24:57.399325 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.772938 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.787780 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.801961 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.818677 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.828109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd6xb\" (UniqueName: \"kubernetes.io/projected/3e576f87-15c8-4a31-b997-5dd3c973e51d-kube-api-access-qd6xb\") pod \"node-resolver-ns7ft\" (UID: \"3e576f87-15c8-4a31-b997-5dd3c973e51d\") " pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.828197 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3e576f87-15c8-4a31-b997-5dd3c973e51d-hosts-file\") pod \"node-resolver-ns7ft\" (UID: \"3e576f87-15c8-4a31-b997-5dd3c973e51d\") " pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.833852 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.847866 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.858209 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.872394 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.884382 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.898612 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:05Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.928884 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3e576f87-15c8-4a31-b997-5dd3c973e51d-hosts-file\") pod \"node-resolver-ns7ft\" (UID: \"3e576f87-15c8-4a31-b997-5dd3c973e51d\") " pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.928926 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd6xb\" (UniqueName: \"kubernetes.io/projected/3e576f87-15c8-4a31-b997-5dd3c973e51d-kube-api-access-qd6xb\") pod \"node-resolver-ns7ft\" (UID: \"3e576f87-15c8-4a31-b997-5dd3c973e51d\") " pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.929314 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3e576f87-15c8-4a31-b997-5dd3c973e51d-hosts-file\") pod \"node-resolver-ns7ft\" (UID: \"3e576f87-15c8-4a31-b997-5dd3c973e51d\") " pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.948395 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd6xb\" (UniqueName: \"kubernetes.io/projected/3e576f87-15c8-4a31-b997-5dd3c973e51d-kube-api-access-qd6xb\") pod \"node-resolver-ns7ft\" (UID: \"3e576f87-15c8-4a31-b997-5dd3c973e51d\") " pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:05 crc kubenswrapper[4796]: I1127 11:25:05.991426 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ns7ft" Nov 27 11:25:06 crc kubenswrapper[4796]: W1127 11:25:06.006558 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e576f87_15c8_4a31_b997_5dd3c973e51d.slice/crio-3000369392f79eda62dc72c6951bffbcf3f79feb0b5863a1a975d4d8ae2eadc4 WatchSource:0}: Error finding container 3000369392f79eda62dc72c6951bffbcf3f79feb0b5863a1a975d4d8ae2eadc4: Status 404 returned error can't find the container with id 3000369392f79eda62dc72c6951bffbcf3f79feb0b5863a1a975d4d8ae2eadc4 Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.046476 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8drhq"] Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.046913 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zfrrg"] Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.047160 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.047880 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.049191 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.049488 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.052569 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.052618 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.052569 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.052943 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.053085 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-v9kkr"] Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.053207 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.053476 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.055801 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.055909 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.056018 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.056043 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.056021 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.073883 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.089370 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.115098 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.132153 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.145725 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.160843 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.179872 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.198706 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.213579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.231950 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-system-cni-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.231989 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1f8355a9-d2c1-4c54-be9d-68ef66397560-cni-binary-copy\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232019 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-netns\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232037 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fa09a212-197e-4df8-9a5f-54321531252b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232057 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-kubelet\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232101 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-socket-dir-parent\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232129 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232174 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjx2d\" (UniqueName: \"kubernetes.io/projected/e606fa06-e313-4bb9-b2cc-84ff65829b3c-kube-api-access-fjx2d\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232216 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-os-release\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232305 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-cni-bin\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232377 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-cni-multus\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232439 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e606fa06-e313-4bb9-b2cc-84ff65829b3c-rootfs\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232459 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-conf-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232492 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-os-release\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232508 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-system-cni-dir\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232523 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-cnibin\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232539 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-cni-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232570 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pnk4\" (UniqueName: \"kubernetes.io/projected/fa09a212-197e-4df8-9a5f-54321531252b-kube-api-access-7pnk4\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232586 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-hostroot\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-daemon-config\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-multus-certs\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232653 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-etc-kubernetes\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232668 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9mjq\" (UniqueName: \"kubernetes.io/projected/1f8355a9-d2c1-4c54-be9d-68ef66397560-kube-api-access-p9mjq\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232693 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e606fa06-e313-4bb9-b2cc-84ff65829b3c-mcd-auth-proxy-config\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232725 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-cnibin\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232742 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-k8s-cni-cncf-io\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232757 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e606fa06-e313-4bb9-b2cc-84ff65829b3c-proxy-tls\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.232850 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fa09a212-197e-4df8-9a5f-54321531252b-cni-binary-copy\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.236579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.252246 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.265597 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.278903 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.293044 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.306894 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.318087 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.333886 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-etc-kubernetes\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.333945 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9mjq\" (UniqueName: \"kubernetes.io/projected/1f8355a9-d2c1-4c54-be9d-68ef66397560-kube-api-access-p9mjq\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.333983 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e606fa06-e313-4bb9-b2cc-84ff65829b3c-mcd-auth-proxy-config\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334006 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-cnibin\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334030 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-k8s-cni-cncf-io\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fa09a212-197e-4df8-9a5f-54321531252b-cni-binary-copy\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334074 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e606fa06-e313-4bb9-b2cc-84ff65829b3c-proxy-tls\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334107 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-netns\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334127 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-system-cni-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334148 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1f8355a9-d2c1-4c54-be9d-68ef66397560-cni-binary-copy\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334171 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fa09a212-197e-4df8-9a5f-54321531252b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334191 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-kubelet\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-socket-dir-parent\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334239 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334294 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e606fa06-e313-4bb9-b2cc-84ff65829b3c-rootfs\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334314 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjx2d\" (UniqueName: \"kubernetes.io/projected/e606fa06-e313-4bb9-b2cc-84ff65829b3c-kube-api-access-fjx2d\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334335 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-os-release\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-cni-bin\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334375 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-cni-multus\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334399 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-conf-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334423 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-os-release\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334446 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-cni-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334470 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-system-cni-dir\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-cnibin\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334512 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-multus-certs\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334533 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pnk4\" (UniqueName: \"kubernetes.io/projected/fa09a212-197e-4df8-9a5f-54321531252b-kube-api-access-7pnk4\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334554 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-hostroot\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.334574 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-daemon-config\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.335432 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-daemon-config\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.335501 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-etc-kubernetes\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.336312 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e606fa06-e313-4bb9-b2cc-84ff65829b3c-mcd-auth-proxy-config\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.336384 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-cnibin\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.336428 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-k8s-cni-cncf-io\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.337009 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fa09a212-197e-4df8-9a5f-54321531252b-cni-binary-copy\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.371531 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-os-release\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.371537 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.371832 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-netns\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.371958 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-system-cni-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.372783 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1f8355a9-d2c1-4c54-be9d-68ef66397560-cni-binary-copy\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.372802 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-socket-dir-parent\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373497 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fa09a212-197e-4df8-9a5f-54321531252b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373514 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e606fa06-e313-4bb9-b2cc-84ff65829b3c-proxy-tls\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373536 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e606fa06-e313-4bb9-b2cc-84ff65829b3c-rootfs\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373566 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-cni-bin\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373606 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373680 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-cni-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373691 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-cnibin\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373714 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-kubelet\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373727 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-var-lib-cni-multus\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373793 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-host-run-multus-certs\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373797 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-hostroot\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373836 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-system-cni-dir\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373869 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1f8355a9-d2c1-4c54-be9d-68ef66397560-multus-conf-dir\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.373869 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fa09a212-197e-4df8-9a5f-54321531252b-os-release\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.374460 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9mjq\" (UniqueName: \"kubernetes.io/projected/1f8355a9-d2c1-4c54-be9d-68ef66397560-kube-api-access-p9mjq\") pod \"multus-8drhq\" (UID: \"1f8355a9-d2c1-4c54-be9d-68ef66397560\") " pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.395645 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pnk4\" (UniqueName: \"kubernetes.io/projected/fa09a212-197e-4df8-9a5f-54321531252b-kube-api-access-7pnk4\") pod \"multus-additional-cni-plugins-zfrrg\" (UID: \"fa09a212-197e-4df8-9a5f-54321531252b\") " pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.400737 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.402135 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjx2d\" (UniqueName: \"kubernetes.io/projected/e606fa06-e313-4bb9-b2cc-84ff65829b3c-kube-api-access-fjx2d\") pod \"machine-config-daemon-v9kkr\" (UID: \"e606fa06-e313-4bb9-b2cc-84ff65829b3c\") " pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.412560 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.418456 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.426038 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6vnft"] Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.426785 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.426918 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.429359 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.429615 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.429878 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.429923 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.430069 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.431073 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.436232 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.436989 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.439919 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.455203 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.469845 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.492889 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.504239 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.517545 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.535956 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-ovn\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536007 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-node-log\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536033 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-netns\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536053 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-var-lib-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536076 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-kubelet\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536098 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-slash\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536118 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-netd\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536136 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-env-overrides\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536156 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-script-lib\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536180 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-systemd\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536206 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-etc-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536235 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536254 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-ovn-kubernetes\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-config\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536324 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-systemd-units\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536315 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536344 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-bin\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536454 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536482 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovn-node-metrics-cert\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536503 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgvrx\" (UniqueName: \"kubernetes.io/projected/9cef17d7-9643-44ec-ae40-cc46b875ed42-kube-api-access-pgvrx\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.536524 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-log-socket\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.550785 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.568339 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:06 crc kubenswrapper[4796]: E1127 11:25:06.568453 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.576596 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.592819 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.607403 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637207 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-systemd-units\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637248 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-bin\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637288 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637311 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovn-node-metrics-cert\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637330 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgvrx\" (UniqueName: \"kubernetes.io/projected/9cef17d7-9643-44ec-ae40-cc46b875ed42-kube-api-access-pgvrx\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637352 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-log-socket\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637390 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-ovn\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637400 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637427 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-bin\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637456 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-node-log\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637380 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-systemd-units\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637414 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-node-log\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637497 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-ovn\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637575 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-netns\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637570 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-log-socket\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637659 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-var-lib-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637628 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-var-lib-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637638 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-netns\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-kubelet\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-slash\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637783 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-netd\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637800 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-env-overrides\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637806 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-kubelet\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637815 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-script-lib\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637846 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-systemd\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637846 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-netd\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637871 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-etc-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637905 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637950 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-ovn-kubernetes\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637956 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-etc-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-config\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.638051 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-systemd\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.638086 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-openvswitch\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.637869 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-slash\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.638129 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-ovn-kubernetes\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.638522 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-env-overrides\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.641664 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-config\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.641817 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovn-node-metrics-cert\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.642044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-script-lib\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.649762 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.664958 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8drhq" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.671431 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgvrx\" (UniqueName: \"kubernetes.io/projected/9cef17d7-9643-44ec-ae40-cc46b875ed42-kube-api-access-pgvrx\") pod \"ovnkube-node-6vnft\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.673924 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" Nov 27 11:25:06 crc kubenswrapper[4796]: W1127 11:25:06.677845 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f8355a9_d2c1_4c54_be9d_68ef66397560.slice/crio-c8b8a21baec1aa48a0e929f40419a23124b71df7d6cd971dc0991d843f35ee8d WatchSource:0}: Error finding container c8b8a21baec1aa48a0e929f40419a23124b71df7d6cd971dc0991d843f35ee8d: Status 404 returned error can't find the container with id c8b8a21baec1aa48a0e929f40419a23124b71df7d6cd971dc0991d843f35ee8d Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.683876 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.700552 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: W1127 11:25:06.704453 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode606fa06_e313_4bb9_b2cc_84ff65829b3c.slice/crio-cdbe07f3a9c8f88672712bbac7019fa62c23dca0f3027a0a101ce5ded84d2058 WatchSource:0}: Error finding container cdbe07f3a9c8f88672712bbac7019fa62c23dca0f3027a0a101ce5ded84d2058: Status 404 returned error can't find the container with id cdbe07f3a9c8f88672712bbac7019fa62c23dca0f3027a0a101ce5ded84d2058 Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.729066 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"cdbe07f3a9c8f88672712bbac7019fa62c23dca0f3027a0a101ce5ded84d2058"} Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.733570 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerStarted","Data":"478e3aca8f184c3c15393c0d563f0401eb6c7347169cf16795750d3c4df20a5d"} Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.736670 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerStarted","Data":"c8b8a21baec1aa48a0e929f40419a23124b71df7d6cd971dc0991d843f35ee8d"} Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.738068 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ns7ft" event={"ID":"3e576f87-15c8-4a31-b997-5dd3c973e51d","Type":"ContainerStarted","Data":"dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec"} Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.738112 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ns7ft" event={"ID":"3e576f87-15c8-4a31-b997-5dd3c973e51d","Type":"ContainerStarted","Data":"3000369392f79eda62dc72c6951bffbcf3f79feb0b5863a1a975d4d8ae2eadc4"} Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.738875 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.740045 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418"} Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.740826 4796 scope.go:117] "RemoveContainer" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:25:06 crc kubenswrapper[4796]: E1127 11:25:06.740940 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.750799 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: W1127 11:25:06.770783 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cef17d7_9643_44ec_ae40_cc46b875ed42.slice/crio-ef1f4ab0b94528f62ddeb26f7105cb62b5c07150876b5e695500d256cd136371 WatchSource:0}: Error finding container ef1f4ab0b94528f62ddeb26f7105cb62b5c07150876b5e695500d256cd136371: Status 404 returned error can't find the container with id ef1f4ab0b94528f62ddeb26f7105cb62b5c07150876b5e695500d256cd136371 Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.783385 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.811623 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.833411 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.844385 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.860926 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.875554 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.890965 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.908076 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.923006 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.933443 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.947359 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.958937 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:06 crc kubenswrapper[4796]: I1127 11:25:06.970284 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:06Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.245302 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.245408 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.245442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245530 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:25:11.245495288 +0000 UTC m=+28.763814216 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245562 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245620 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:11.245602591 +0000 UTC m=+28.763921579 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.245644 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245653 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245707 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245713 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245723 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245744 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:11.245735854 +0000 UTC m=+28.764054783 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.245803 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:11.245779556 +0000 UTC m=+28.764098474 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.346196 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.346410 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.346441 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.346457 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.346565 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:11.346542421 +0000 UTC m=+28.864861359 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.548546 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.555320 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.562152 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.568893 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.568979 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.569062 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.569187 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.571727 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.598874 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.625752 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.638320 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.649459 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.659468 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.672656 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.687579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.704117 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.715693 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.729903 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.745958 4796 generic.go:334] "Generic (PLEG): container finished" podID="fa09a212-197e-4df8-9a5f-54321531252b" containerID="afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16" exitCode=0 Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.746083 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerDied","Data":"afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16"} Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.748722 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerStarted","Data":"7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663"} Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.751695 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934"} Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.751724 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e"} Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.754168 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.762931 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" exitCode=0 Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.763016 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.763087 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"ef1f4ab0b94528f62ddeb26f7105cb62b5c07150876b5e695500d256cd136371"} Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.771029 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: E1127 11:25:07.771073 4796 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.788752 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.813514 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.826387 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.848631 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.869118 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.882965 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.898289 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.912720 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.925628 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.937595 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.953902 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.968568 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:07 crc kubenswrapper[4796]: I1127 11:25:07.982602 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.000499 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:07Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.568912 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:08 crc kubenswrapper[4796]: E1127 11:25:08.569043 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.768775 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.768826 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.768835 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.768846 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.771286 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerStarted","Data":"33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0"} Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.783443 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.796976 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.813999 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.826038 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.837250 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.856240 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.889617 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.913313 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.925455 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.938492 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.953673 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.963839 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.976332 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:08 crc kubenswrapper[4796]: I1127 11:25:08.990719 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:08Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.569023 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.569101 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.569615 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.569426 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.790008 4796 generic.go:334] "Generic (PLEG): container finished" podID="fa09a212-197e-4df8-9a5f-54321531252b" containerID="33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0" exitCode=0 Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.790085 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerDied","Data":"33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.796030 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.796087 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.799821 4796 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.802388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.802442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.802462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.802600 4796 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.810750 4796 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.811072 4796 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.812329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.812401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.812421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.812445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.812474 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:09Z","lastTransitionTime":"2025-11-27T11:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.815851 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.836230 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.844742 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.848579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.850729 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.850768 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.850785 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.850809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.850826 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:09Z","lastTransitionTime":"2025-11-27T11:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.860885 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.876191 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.876519 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.882516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.882558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.882569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.882589 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.882602 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:09Z","lastTransitionTime":"2025-11-27T11:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.899005 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.901054 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.910473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.910526 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.910537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.910556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.910568 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:09Z","lastTransitionTime":"2025-11-27T11:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.925605 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.927811 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.932919 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.932956 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.932968 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.932988 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.933001 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:09Z","lastTransitionTime":"2025-11-27T11:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.943743 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.948833 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: E1127 11:25:09.948952 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.950505 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.950524 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.950532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.950545 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.950554 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:09Z","lastTransitionTime":"2025-11-27T11:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.960610 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.978035 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:09 crc kubenswrapper[4796]: I1127 11:25:09.992539 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:09Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.008317 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.024082 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.041563 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.053317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.053360 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.053373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.053393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.053407 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.155590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.155618 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.155630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.155644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.155653 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.258423 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.258471 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.258484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.258503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.258516 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.361026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.361260 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.361290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.361307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.361320 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.463420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.463484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.463501 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.463525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.463543 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.566582 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.566619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.566628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.566644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.566654 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.568757 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:10 crc kubenswrapper[4796]: E1127 11:25:10.568881 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.668783 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.668850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.668875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.668904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.668968 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.772243 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.772311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.772324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.772339 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.772351 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.802900 4796 generic.go:334] "Generic (PLEG): container finished" podID="fa09a212-197e-4df8-9a5f-54321531252b" containerID="02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f" exitCode=0 Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.802963 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerDied","Data":"02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.837544 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.856920 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.875238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.875328 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.875346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.875369 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.875396 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.876565 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.895682 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.909759 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.934063 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.949991 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.964564 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.978939 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.979024 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.979046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.979079 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.979105 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:10Z","lastTransitionTime":"2025-11-27T11:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:10 crc kubenswrapper[4796]: I1127 11:25:10.983842 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.000462 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:10Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.019772 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.036587 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.059315 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.072625 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.081919 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.081979 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.082001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.082033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.082056 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.184327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.184366 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.184424 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.184439 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.184451 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.284127 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.284256 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.284325 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284400 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:25:19.284361394 +0000 UTC m=+36.802680362 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284454 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284497 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:19.284487478 +0000 UTC m=+36.802806386 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284712 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284771 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:19.284754425 +0000 UTC m=+36.803073383 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.284450 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284908 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284942 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.284961 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.285006 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:19.284990321 +0000 UTC m=+36.803309269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.286882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.286945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.286973 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.287009 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.287030 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.385738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.386031 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.386061 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.386081 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.386148 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:19.386125816 +0000 UTC m=+36.904444774 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.423311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.423365 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.423383 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.423405 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.423422 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.526502 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.526579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.526662 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.526712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.526756 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.568374 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.568469 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.568591 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:11 crc kubenswrapper[4796]: E1127 11:25:11.568686 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.629731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.629796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.629812 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.629831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.629843 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.732732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.732793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.732807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.732830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.732844 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.808797 4796 generic.go:334] "Generic (PLEG): container finished" podID="fa09a212-197e-4df8-9a5f-54321531252b" containerID="fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2" exitCode=0 Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.808840 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerDied","Data":"fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.824744 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.835755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.835791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.835802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.835821 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.835838 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.838971 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.856851 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.871308 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.883070 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.896602 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.907799 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.923840 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.936552 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.941811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.941836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.941845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.941858 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.941867 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:11Z","lastTransitionTime":"2025-11-27T11:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.950434 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.972495 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.983695 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:11 crc kubenswrapper[4796]: I1127 11:25:11.995859 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.009796 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.044852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.044894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.044907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.044926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.044942 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.146928 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.146957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.146965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.146978 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.146987 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.249611 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.249678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.249690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.249706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.249719 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.352038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.352080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.352088 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.352102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.352113 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.455097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.455145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.455162 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.455185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.455204 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.457138 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-4z4td"] Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.457781 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.461749 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.463643 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.464118 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.465564 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.480151 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.494887 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.528596 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.544105 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.556348 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.557906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.557957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.557969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.557987 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.557998 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.568415 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:12 crc kubenswrapper[4796]: E1127 11:25:12.568799 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.593581 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.598390 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4e236427-8ad0-41e6-861c-ed26c11c80fc-serviceca\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.598524 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4e236427-8ad0-41e6-861c-ed26c11c80fc-host\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.598667 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w9pl\" (UniqueName: \"kubernetes.io/projected/4e236427-8ad0-41e6-861c-ed26c11c80fc-kube-api-access-5w9pl\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.608369 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.622244 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.633804 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.647898 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.660473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.660762 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.660853 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.660923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.661005 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.668812 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.683209 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.698093 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.699442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4e236427-8ad0-41e6-861c-ed26c11c80fc-serviceca\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.699578 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4e236427-8ad0-41e6-861c-ed26c11c80fc-host\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.699771 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w9pl\" (UniqueName: \"kubernetes.io/projected/4e236427-8ad0-41e6-861c-ed26c11c80fc-kube-api-access-5w9pl\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.699872 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4e236427-8ad0-41e6-861c-ed26c11c80fc-host\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.700611 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4e236427-8ad0-41e6-861c-ed26c11c80fc-serviceca\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.721206 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.727119 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w9pl\" (UniqueName: \"kubernetes.io/projected/4e236427-8ad0-41e6-861c-ed26c11c80fc-kube-api-access-5w9pl\") pod \"node-ca-4z4td\" (UID: \"4e236427-8ad0-41e6-861c-ed26c11c80fc\") " pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.739970 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.763627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.763674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.763685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.763700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.763714 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.774916 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4z4td" Nov 27 11:25:12 crc kubenswrapper[4796]: W1127 11:25:12.789331 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e236427_8ad0_41e6_861c_ed26c11c80fc.slice/crio-8d346924e39a8438ae06ef57fa668bac438b9806de1687d8aa5e12dd8e1f8fa5 WatchSource:0}: Error finding container 8d346924e39a8438ae06ef57fa668bac438b9806de1687d8aa5e12dd8e1f8fa5: Status 404 returned error can't find the container with id 8d346924e39a8438ae06ef57fa668bac438b9806de1687d8aa5e12dd8e1f8fa5 Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.813257 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4z4td" event={"ID":"4e236427-8ad0-41e6-861c-ed26c11c80fc","Type":"ContainerStarted","Data":"8d346924e39a8438ae06ef57fa668bac438b9806de1687d8aa5e12dd8e1f8fa5"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.818597 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerStarted","Data":"092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.832936 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.844943 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.864775 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.867809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.867846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.867859 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.867904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.867918 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.880203 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.897762 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.910383 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.942478 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.960329 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.970576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.970605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.970616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.970631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.970643 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:12Z","lastTransitionTime":"2025-11-27T11:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.973284 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.985969 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:12 crc kubenswrapper[4796]: I1127 11:25:12.997469 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:12Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.020448 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.036553 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.054980 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.072609 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.072636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.072643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.072656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.072665 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.081783 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.095578 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.175973 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.176032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.176048 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.176082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.176100 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.278832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.278886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.278901 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.278930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.278945 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.382322 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.382382 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.382397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.382421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.382433 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.485672 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.485728 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.485747 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.485773 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.485793 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.568958 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.568963 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:13 crc kubenswrapper[4796]: E1127 11:25:13.569082 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:13 crc kubenswrapper[4796]: E1127 11:25:13.569185 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.588646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.588732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.588758 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.588789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.588815 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.592474 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.610463 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.628892 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.645362 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.662670 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.690497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.690532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.690542 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.690558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.690569 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.693514 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.710193 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.743109 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.758687 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.773020 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.789435 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.793241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.793335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.793350 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.793371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.793385 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.802910 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.816927 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.830479 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.838370 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4z4td" event={"ID":"4e236427-8ad0-41e6-861c-ed26c11c80fc","Type":"ContainerStarted","Data":"ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.842081 4796 generic.go:334] "Generic (PLEG): container finished" podID="fa09a212-197e-4df8-9a5f-54321531252b" containerID="092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594" exitCode=0 Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.842130 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerDied","Data":"092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.844558 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.857753 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.874888 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.891837 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.903984 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.908777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.908865 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.908933 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.909004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.909062 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:13Z","lastTransitionTime":"2025-11-27T11:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.916149 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.933074 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.945450 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.970139 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:13 crc kubenswrapper[4796]: I1127 11:25:13.998196 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.011514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.011541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.011550 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.011564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.011574 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.021024 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:14Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.037343 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:14Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.051890 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:14Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.069554 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:14Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.088779 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:14Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.107263 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:14Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.114661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.114871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.115003 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.115186 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.115376 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.219439 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.219507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.219524 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.219554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.219593 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.328365 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.328448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.328470 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.328493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.328511 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.431964 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.432053 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.432076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.432113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.432137 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.535457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.535516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.535533 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.535556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.535576 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.568560 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:14 crc kubenswrapper[4796]: E1127 11:25:14.568737 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.638203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.638235 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.638245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.638260 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.638287 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.740594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.740856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.740922 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.740990 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.741045 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.844622 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.844786 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.844803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.844824 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.844840 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.851229 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerStarted","Data":"bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186"} Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.947801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.947837 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.947846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.947858 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:14 crc kubenswrapper[4796]: I1127 11:25:14.947867 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:14Z","lastTransitionTime":"2025-11-27T11:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.050756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.050822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.050890 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.050923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.050945 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.154018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.154102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.154119 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.154144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.154162 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.256705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.256772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.256795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.256825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.256850 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.359934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.360187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.360199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.360214 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.360226 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.463699 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.463753 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.463776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.463801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.463818 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.567321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.567373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.567389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.567411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.567428 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.568247 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:15 crc kubenswrapper[4796]: E1127 11:25:15.568438 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.568716 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:15 crc kubenswrapper[4796]: E1127 11:25:15.568882 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.671090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.671151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.671168 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.671193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.671213 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.773767 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.774542 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.774578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.774610 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.774629 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.877169 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.877247 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.877303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.877333 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.877353 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.980225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.980327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.980354 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.980387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:15 crc kubenswrapper[4796]: I1127 11:25:15.980409 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:15Z","lastTransitionTime":"2025-11-27T11:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.083449 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.083495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.083511 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.083534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.083551 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.186822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.186899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.186923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.186957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.186981 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.290525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.290623 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.290645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.290668 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.290685 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.394842 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.394900 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.394916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.394940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.394960 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.498240 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.498346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.498370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.498399 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.498424 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.568859 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:16 crc kubenswrapper[4796]: E1127 11:25:16.569033 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.606093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.606563 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.606787 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.606835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.606864 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.710390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.710478 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.710510 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.710539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.710560 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.813808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.813870 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.813889 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.813914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.813932 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.917156 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.917235 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.917255 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.917304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:16 crc kubenswrapper[4796]: I1127 11:25:16.917324 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:16Z","lastTransitionTime":"2025-11-27T11:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.020461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.020551 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.020575 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.020603 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.020625 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.124116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.124173 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.124209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.124236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.124256 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.227906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.228002 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.228059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.228082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.228100 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.331051 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.331102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.331119 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.331144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.331160 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.433528 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.433592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.433617 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.433648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.433669 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.536914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.536985 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.537010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.537037 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.537057 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.568841 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.568841 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:17 crc kubenswrapper[4796]: E1127 11:25:17.569017 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:17 crc kubenswrapper[4796]: E1127 11:25:17.569164 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.640004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.640060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.640079 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.640101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.640121 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.743883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.743956 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.743980 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.744012 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.744036 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.846851 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.846906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.846923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.846945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.846962 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.875341 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b"} Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.949318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.950183 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.950394 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.950536 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:17 crc kubenswrapper[4796]: I1127 11:25:17.950665 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:17Z","lastTransitionTime":"2025-11-27T11:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.053962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.054024 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.054046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.054077 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.054097 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.156543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.156597 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.156615 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.156640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.156657 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.260089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.260162 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.260181 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.260206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.260224 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.363313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.363648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.363776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.363993 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.364180 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.467695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.467931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.468068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.468239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.468487 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.561610 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5"] Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.562356 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.567298 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.567498 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.568124 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:18 crc kubenswrapper[4796]: E1127 11:25:18.568474 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.571676 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.571925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.571942 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.571965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.571985 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.590206 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.610780 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.628907 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.663174 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.669617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88565635-1352-4c89-9a97-7cbe728d543e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.669709 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88565635-1352-4c89-9a97-7cbe728d543e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.669761 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqhv\" (UniqueName: \"kubernetes.io/projected/88565635-1352-4c89-9a97-7cbe728d543e-kube-api-access-zsqhv\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.669825 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88565635-1352-4c89-9a97-7cbe728d543e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.674581 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.674629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.674639 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.674655 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.674668 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.681453 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.698107 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.716525 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.732557 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.765243 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.771153 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88565635-1352-4c89-9a97-7cbe728d543e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.771230 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88565635-1352-4c89-9a97-7cbe728d543e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.771361 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88565635-1352-4c89-9a97-7cbe728d543e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.771432 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqhv\" (UniqueName: \"kubernetes.io/projected/88565635-1352-4c89-9a97-7cbe728d543e-kube-api-access-zsqhv\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.772811 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88565635-1352-4c89-9a97-7cbe728d543e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.773061 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88565635-1352-4c89-9a97-7cbe728d543e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.778820 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.778895 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.778920 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.778950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.778973 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.780225 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88565635-1352-4c89-9a97-7cbe728d543e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.787534 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.801719 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqhv\" (UniqueName: \"kubernetes.io/projected/88565635-1352-4c89-9a97-7cbe728d543e-kube-api-access-zsqhv\") pod \"ovnkube-control-plane-749d76644c-jngs5\" (UID: \"88565635-1352-4c89-9a97-7cbe728d543e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.807720 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.831000 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.854482 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.873412 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.881910 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.881966 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.881983 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.882004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.882024 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.884824 4796 generic.go:334] "Generic (PLEG): container finished" podID="fa09a212-197e-4df8-9a5f-54321531252b" containerID="bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186" exitCode=0 Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.884917 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerDied","Data":"bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.894900 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.895429 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.916647 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: W1127 11:25:18.918502 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88565635_1352_4c89_9a97_7cbe728d543e.slice/crio-16e05f380f0262d322f14a81b8e49969d75dd20927341cc4e331ea27b8548884 WatchSource:0}: Error finding container 16e05f380f0262d322f14a81b8e49969d75dd20927341cc4e331ea27b8548884: Status 404 returned error can't find the container with id 16e05f380f0262d322f14a81b8e49969d75dd20927341cc4e331ea27b8548884 Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.942013 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.965320 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.985578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.985635 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.985658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.985686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.985711 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:18Z","lastTransitionTime":"2025-11-27T11:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:18 crc kubenswrapper[4796]: I1127 11:25:18.986186 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:18Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.051453 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.064829 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.088298 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.088346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.088358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.088374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.088385 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.089030 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.103110 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.118092 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.130572 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.142824 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.156705 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.176810 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.190867 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.191162 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.191291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.191302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.191317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.191328 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.203384 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.223426 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.240625 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.294547 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.294914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.295114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.295256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.295443 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.332218 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-79ll4"] Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.336057 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.336231 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.375627 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.378869 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.378991 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.379029 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379105 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:25:35.379072977 +0000 UTC m=+52.897391905 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379126 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379145 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379155 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379154 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.379192 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379212 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:35.37919613 +0000 UTC m=+52.897515048 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379229 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:35.379222421 +0000 UTC m=+52.897541339 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379339 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.379373 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:35.379366995 +0000 UTC m=+52.897685913 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.398093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.398176 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.398206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.398239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.398305 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.399737 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.423258 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.440642 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.452827 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.481052 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcv6f\" (UniqueName: \"kubernetes.io/projected/0494bee3-7923-49de-8c4f-e0fa4ffad936-kube-api-access-tcv6f\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.481144 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.481309 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.481541 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.481577 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.481600 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.481682 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:35.481653661 +0000 UTC m=+52.999972619 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.485554 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.497733 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.501887 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.501926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.501934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.501949 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.501960 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.517579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.533416 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.551003 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.565882 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.568227 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.568243 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.568409 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.568769 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.581801 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcv6f\" (UniqueName: \"kubernetes.io/projected/0494bee3-7923-49de-8c4f-e0fa4ffad936-kube-api-access-tcv6f\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.581845 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.581980 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: E1127 11:25:19.582030 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:20.082015865 +0000 UTC m=+37.600334793 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.588519 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.604168 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcv6f\" (UniqueName: \"kubernetes.io/projected/0494bee3-7923-49de-8c4f-e0fa4ffad936-kube-api-access-tcv6f\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.604735 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.604781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.604793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.604810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.604825 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.605240 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.619901 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.637479 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.655069 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.669989 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.707466 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.707498 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.707511 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.707528 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.707540 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.812053 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.812102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.812119 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.812141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.812157 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.888927 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" event={"ID":"88565635-1352-4c89-9a97-7cbe728d543e","Type":"ContainerStarted","Data":"16e05f380f0262d322f14a81b8e49969d75dd20927341cc4e331ea27b8548884"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.906421 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.918104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.918171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.918188 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.918211 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.918229 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:19Z","lastTransitionTime":"2025-11-27T11:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.919774 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.937317 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.950506 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.980332 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:19 crc kubenswrapper[4796]: I1127 11:25:19.999557 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:19Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.015251 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.020512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.020605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.020619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.020635 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.020644 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.034119 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.047702 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.062580 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.075431 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.087680 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.089745 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.089921 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.089977 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:21.089960001 +0000 UTC m=+38.608278919 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.101429 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.120913 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.122165 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.122212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.122226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.122242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.122253 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.137983 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.140960 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.141013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.141031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.141050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.141069 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.153725 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.156165 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.157030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.157068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.157082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.157097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.157109 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.167743 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.172960 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.176528 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.176590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.176608 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.176631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.176650 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.190320 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.194474 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.194524 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.194540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.194561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.194577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.212699 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.216706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.216745 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.216761 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.216784 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.216800 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.234077 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:20Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.234327 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.236603 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.236843 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.236861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.236883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.236901 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.340460 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.340505 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.340516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.340534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.340552 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.444398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.444480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.444503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.444534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.444561 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.547166 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.547569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.547706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.547830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.547945 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.568217 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:20 crc kubenswrapper[4796]: E1127 11:25:20.568568 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.655018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.655091 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.655117 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.655147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.655180 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.758632 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.759134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.759321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.759484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.759622 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.863026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.863069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.863080 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.863097 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.863108 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.894480 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" event={"ID":"88565635-1352-4c89-9a97-7cbe728d543e","Type":"ContainerStarted","Data":"bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad"} Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.966087 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.966157 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.966179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.966206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:20 crc kubenswrapper[4796]: I1127 11:25:20.966229 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:20Z","lastTransitionTime":"2025-11-27T11:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.069987 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.070057 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.070075 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.070100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.070118 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.101031 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:21 crc kubenswrapper[4796]: E1127 11:25:21.101348 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:21 crc kubenswrapper[4796]: E1127 11:25:21.101465 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:23.101431732 +0000 UTC m=+40.619750690 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.173304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.173370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.173390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.173417 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.173434 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.276705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.276792 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.276810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.276832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.276850 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.379950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.380013 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.380034 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.380062 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.380096 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.483189 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.483262 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.483320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.483348 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.483366 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.568910 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.568910 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.569176 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:21 crc kubenswrapper[4796]: E1127 11:25:21.569443 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:21 crc kubenswrapper[4796]: E1127 11:25:21.569559 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.569666 4796 scope.go:117] "RemoveContainer" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:25:21 crc kubenswrapper[4796]: E1127 11:25:21.569688 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.586089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.586165 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.586359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.586469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.586501 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.690138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.690198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.690217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.690241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.690258 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.793914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.794245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.794454 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.794587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.794893 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.898050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.898109 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.898127 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.898152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.898170 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:21Z","lastTransitionTime":"2025-11-27T11:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:21 crc kubenswrapper[4796]: I1127 11:25:21.904374 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" event={"ID":"fa09a212-197e-4df8-9a5f-54321531252b","Type":"ContainerStarted","Data":"59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.001098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.001165 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.001183 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.001207 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.001225 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.104506 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.104568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.104585 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.104614 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.104639 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.207759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.207833 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.207856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.207905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.207928 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.310806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.310847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.310864 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.310886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.310903 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.414049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.414151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.414170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.414202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.414221 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.517327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.517407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.517432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.517467 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.517493 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.568092 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:22 crc kubenswrapper[4796]: E1127 11:25:22.568229 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.619900 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.619953 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.619968 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.619987 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.620329 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.723592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.723705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.723730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.723762 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.723780 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.831654 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.831702 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.831716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.831737 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.831747 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.935109 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.935199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.935220 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.935252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:22 crc kubenswrapper[4796]: I1127 11:25:22.935317 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:22Z","lastTransitionTime":"2025-11-27T11:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.038459 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.038567 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.038590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.038620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.038645 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.128150 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:23 crc kubenswrapper[4796]: E1127 11:25:23.129208 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:23 crc kubenswrapper[4796]: E1127 11:25:23.129604 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:27.129453853 +0000 UTC m=+44.647772921 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.141792 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.141867 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.141889 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.141923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.141944 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.244946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.245018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.245044 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.245077 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.245098 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.348057 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.348110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.348124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.348146 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.348161 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.453839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.453886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.453902 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.453922 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.453933 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.556170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.556236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.556254 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.556311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.556336 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.568793 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.568875 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:23 crc kubenswrapper[4796]: E1127 11:25:23.569117 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.569173 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:23 crc kubenswrapper[4796]: E1127 11:25:23.569329 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:23 crc kubenswrapper[4796]: E1127 11:25:23.569441 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.588104 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.608013 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.628734 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.645823 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.658633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.658667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.658675 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.658689 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.658699 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.663621 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.677835 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.741509 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.760495 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.761084 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.761145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.761163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.761190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.761207 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.778004 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.796688 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.823476 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.846192 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.864147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.864494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.864836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.865036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.865178 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.866331 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.881590 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.894551 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.915830 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.927261 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.968177 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.968496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.968643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.968784 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:23 crc kubenswrapper[4796]: I1127 11:25:23.968951 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:23Z","lastTransitionTime":"2025-11-27T11:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.071483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.071516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.071527 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.071543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.071555 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.173480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.173519 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.173529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.173545 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.173558 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.277254 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.277354 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.277377 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.277401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.277422 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.379647 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.379685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.379697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.379713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.379726 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.483776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.483839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.483858 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.483882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.483901 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.568234 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:24 crc kubenswrapper[4796]: E1127 11:25:24.568417 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.586598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.586811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.587069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.587340 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.587543 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.690886 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.690941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.690957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.690976 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.690991 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.793507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.793553 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.793567 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.793585 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.793600 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.896549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.896594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.896606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.896622 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.896633 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:24Z","lastTransitionTime":"2025-11-27T11:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.930370 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:24Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.942557 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:24Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.963299 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:24Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:24 crc kubenswrapper[4796]: I1127 11:25:24.974992 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:24Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.000545 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.000587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.000601 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.000620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.000633 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.003045 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:24Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.021135 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.039991 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.061976 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.078470 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.097588 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.103116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.103145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.103160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.103179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.103194 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.119567 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.134579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.154208 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.170229 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.188691 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.201868 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.205972 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.206023 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.206034 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.206050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.206084 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.222424 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:25Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.309159 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.309234 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.309256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.309321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.309360 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.412810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.412879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.412901 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.412931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.412956 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.515990 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.516052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.516069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.516092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.516109 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.568210 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:25 crc kubenswrapper[4796]: E1127 11:25:25.568502 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.569200 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:25 crc kubenswrapper[4796]: E1127 11:25:25.569386 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.569504 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:25 crc kubenswrapper[4796]: E1127 11:25:25.569626 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.618830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.618888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.618908 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.618932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.618950 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.722425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.722482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.722502 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.722525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.722542 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.826543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.826643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.826662 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.826723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:25 crc kubenswrapper[4796]: I1127 11:25:25.826742 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:25Z","lastTransitionTime":"2025-11-27T11:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.320912 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.320970 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.320989 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.321012 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.321030 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.326597 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" event={"ID":"88565635-1352-4c89-9a97-7cbe728d543e","Type":"ContainerStarted","Data":"41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.424347 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.424398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.424415 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.424437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.424456 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.528628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.529056 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.529081 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.529105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.529124 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.568775 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:26 crc kubenswrapper[4796]: E1127 11:25:26.568989 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.632776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.632832 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.632845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.632863 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.632874 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.735808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.735857 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.735868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.735883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.735969 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.739465 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.739496 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.739617 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.812076 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.813142 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.841518 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.841742 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.841797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.841816 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.841843 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.841861 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.861235 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.873599 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.888517 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.923564 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.946195 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.946369 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.946422 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.946443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.946470 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.946498 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:26Z","lastTransitionTime":"2025-11-27T11:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.967034 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:26 crc kubenswrapper[4796]: I1127 11:25:26.983671 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.000395 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:26Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.036161 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.051776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.051827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.051844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.051866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.051887 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.061324 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.085379 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.110187 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.125171 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.141223 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.154984 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.155027 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.155042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.155062 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.155079 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.160693 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.172828 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.190141 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.206845 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.221570 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.227033 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:27 crc kubenswrapper[4796]: E1127 11:25:27.227165 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:27 crc kubenswrapper[4796]: E1127 11:25:27.227208 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:35.227195843 +0000 UTC m=+52.745514761 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.231585 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.247419 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.256983 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.257008 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.257016 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.257028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.257037 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.259716 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.271160 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.281119 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.288545 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.303068 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.313224 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.326012 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.330812 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.332342 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.332468 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.344353 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.359629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.359667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.359678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.359694 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.359706 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.360252 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.372846 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.384994 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.398359 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.412253 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.426492 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.439514 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.456759 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.461598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.461640 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.461655 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.461678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.461693 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.473808 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.487262 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.509241 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.527395 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.540091 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.552351 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.565956 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.566033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.566060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.566090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.566113 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.568223 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:27 crc kubenswrapper[4796]: E1127 11:25:27.568380 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.568749 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:27 crc kubenswrapper[4796]: E1127 11:25:27.568814 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.568932 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:27 crc kubenswrapper[4796]: E1127 11:25:27.568984 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.568872 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.599842 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.625883 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.648046 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.659787 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.668351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.668398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.668409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.668425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.668433 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.676167 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.685470 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:27Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.771422 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.771458 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.771467 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.771480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.771493 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.874136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.874193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.874202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.874214 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.874245 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.977397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.977457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.977473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.977503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:27 crc kubenswrapper[4796]: I1127 11:25:27.977526 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:27Z","lastTransitionTime":"2025-11-27T11:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.081655 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.081995 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.082187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.082373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.082508 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.185708 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.186059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.186509 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.186818 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.187217 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.290636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.290685 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.290701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.290725 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.290743 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.337578 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.338522 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.361711 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.394086 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.394180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.394205 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.394234 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.394257 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.400815 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.420814 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.440624 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.463831 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.484324 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.498139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.498206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.498224 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.498303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.498330 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.505307 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.524347 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.547021 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.569332 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:28 crc kubenswrapper[4796]: E1127 11:25:28.569576 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.580762 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.601061 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.601130 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.601149 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.601174 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.601192 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.602421 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.640498 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.661517 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.680463 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.701212 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.703170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.703225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.703243 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.703299 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.703317 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.713984 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.734006 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:28Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.806093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.806163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.806182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.806208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.806226 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.909096 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.909218 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.909237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.909291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:28 crc kubenswrapper[4796]: I1127 11:25:28.909311 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:28Z","lastTransitionTime":"2025-11-27T11:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.012251 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.012364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.012382 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.012407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.012427 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.115444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.115500 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.115517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.115539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.115556 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.219158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.219216 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.219236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.219306 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.219343 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.322004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.322045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.322061 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.322083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.322099 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.425428 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.425490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.425509 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.425534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.425553 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.528878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.528937 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.528955 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.528974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.528986 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.568779 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.568856 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:29 crc kubenswrapper[4796]: E1127 11:25:29.568909 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.568952 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:29 crc kubenswrapper[4796]: E1127 11:25:29.569025 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:29 crc kubenswrapper[4796]: E1127 11:25:29.569148 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.630999 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.631059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.631071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.631086 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.631098 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.739860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.739915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.739933 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.739955 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.739973 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.842948 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.843010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.843040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.843066 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.843084 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.946225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.946297 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.946316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.946335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:29 crc kubenswrapper[4796]: I1127 11:25:29.946349 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:29Z","lastTransitionTime":"2025-11-27T11:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.049749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.049849 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.049871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.049897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.049923 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.152835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.152909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.152926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.152952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.152971 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.255882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.255933 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.255952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.255976 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.255993 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.345518 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.345583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.345605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.345633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.345654 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.367354 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:30Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.372533 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.372624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.372648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.372677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.372700 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.393922 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:30Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.398425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.398490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.398514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.398541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.398562 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.420870 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:30Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.425493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.425546 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.425564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.425587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.425606 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.446541 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:30Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.452219 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.452295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.452313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.452393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.452412 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.472998 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:30Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.473218 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.475389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.475474 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.475491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.475514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.475536 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.569129 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:30 crc kubenswrapper[4796]: E1127 11:25:30.569402 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.577991 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.578044 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.578060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.578128 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.578160 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.681595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.681667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.681691 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.681722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.681745 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.785043 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.785114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.785133 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.785163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.785187 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.888120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.888167 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.888179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.888196 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.888210 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.991373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.991431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.991441 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.991460 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:30 crc kubenswrapper[4796]: I1127 11:25:30.991474 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:30Z","lastTransitionTime":"2025-11-27T11:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.095135 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.095199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.095210 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.095228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.095240 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.198752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.198829 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.198847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.198874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.198896 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.302671 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.302747 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.302765 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.302790 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.302809 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.406075 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.406152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.406171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.406200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.406220 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.511962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.512036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.512079 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.512110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.512134 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.568427 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.568561 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.568427 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:31 crc kubenswrapper[4796]: E1127 11:25:31.568906 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:31 crc kubenswrapper[4796]: E1127 11:25:31.569059 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:31 crc kubenswrapper[4796]: E1127 11:25:31.568709 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.615326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.615366 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.615375 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.615388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.615398 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.718063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.718141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.718195 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.718225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.718246 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.821234 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.821359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.821381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.821407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.821427 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.925161 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.925230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.925252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.925315 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:31 crc kubenswrapper[4796]: I1127 11:25:31.925341 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:31Z","lastTransitionTime":"2025-11-27T11:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.028855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.028916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.028933 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.028957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.028975 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.132353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.132411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.132431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.132456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.132472 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.235426 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.235480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.235497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.235517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.235533 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.338963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.339045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.339070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.339099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.339126 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.442361 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.442407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.442422 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.442443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.442460 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.546106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.546159 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.546172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.546189 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.546205 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.568686 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:32 crc kubenswrapper[4796]: E1127 11:25:32.568889 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.649587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.649635 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.649644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.649658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.649670 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.752576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.752607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.752617 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.752631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.752642 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.855814 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.855878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.855896 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.855921 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.855958 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.958752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.958827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.958848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.958872 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:32 crc kubenswrapper[4796]: I1127 11:25:32.958897 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:32Z","lastTransitionTime":"2025-11-27T11:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.062021 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.062086 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.062104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.062128 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.062145 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.165606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.165681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.165699 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.165726 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.165746 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.269123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.269179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.269197 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.269232 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.269249 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.371222 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.371715 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.371728 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.371744 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.371757 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.475369 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.475429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.475445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.475469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.475486 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.568785 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.568802 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.569044 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.569072 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.569235 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.569419 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.577449 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.577491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.577503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.577516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.577529 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.595682 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.614277 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.630795 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.646370 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.660654 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.683764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.683810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.683830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.683853 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.683871 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.694676 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.720688 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.753096 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.770589 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.787722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.787794 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.787811 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.787837 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.787855 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.789315 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.816091 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.834348 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.854626 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.876355 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.890592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.890661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.890681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.890705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.890725 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.895408 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.895587 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.895807 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.896022 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.897106 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.897566 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.897632 4796 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.898082 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.898443 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.898785 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 27 11:25:33 crc kubenswrapper[4796]: E1127 11:25:33.898822 4796 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.943663 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.963214 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:33Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.994314 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.994350 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.994366 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.994389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:33 crc kubenswrapper[4796]: I1127 11:25:33.994409 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:33Z","lastTransitionTime":"2025-11-27T11:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.098070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.098190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.098215 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.098246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.098301 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.201418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.201484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.201501 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.201530 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.201548 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.304962 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.305007 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.305434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.305462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.305474 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.366358 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/0.log" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.370609 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" exitCode=1 Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.370660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.371626 4796 scope.go:117] "RemoveContainer" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.392968 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.409854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.409907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.409926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.409952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.409971 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.418301 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.436371 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.460712 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.481511 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.498733 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.513941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.513989 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.513997 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.514014 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.514024 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.533425 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:33Z\\\",\\\"message\\\":\\\"Sending *v1.Pod event handler 6 for removal\\\\nI1127 11:25:33.705752 6057 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 11:25:33.705779 6057 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 11:25:33.705796 6057 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 11:25:33.705810 6057 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 11:25:33.705819 6057 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 11:25:33.705421 6057 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:25:33.705483 6057 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:25:33.707013 6057 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 11:25:33.707049 6057 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 11:25:33.707071 6057 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:25:33.707158 6057 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:25:33.707983 6057 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.548171 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.569145 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:34 crc kubenswrapper[4796]: E1127 11:25:34.569446 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.585057 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.602660 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.616939 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.617014 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.617063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.617090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.617109 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.623292 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.650416 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.669565 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.697622 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.713991 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.719115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.719201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.719219 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.719242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.719259 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.729408 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.743571 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:34Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.822565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.822620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.822636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.822658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.822670 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.925723 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.925783 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.925801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.925828 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:34 crc kubenswrapper[4796]: I1127 11:25:34.925845 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:34Z","lastTransitionTime":"2025-11-27T11:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.029650 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.030071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.030091 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.030113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.030130 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.132366 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.132410 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.132421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.132438 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.132450 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.235125 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.235167 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.235178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.235192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.235203 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.321254 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.321487 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.321602 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:25:51.321573257 +0000 UTC m=+68.839892215 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.337774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.337828 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.337846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.337867 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.337881 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.377131 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/0.log" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.380408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.380755 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.399886 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.420702 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.421937 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.422061 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.422139 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.422214 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422245 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422371 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422469 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422514 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422540 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422259 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:26:07.422209208 +0000 UTC m=+84.940528166 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422764 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:26:07.422705271 +0000 UTC m=+84.941024229 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422829 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:26:07.422811084 +0000 UTC m=+84.941130042 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.422863 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:26:07.422851565 +0000 UTC m=+84.941170523 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.440845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.440911 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.440930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.440955 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.440974 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.452722 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.474027 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.494369 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.508440 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.527573 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.527772 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.527987 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.528051 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.528074 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.528174 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:26:07.528137633 +0000 UTC m=+85.046456591 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.544805 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.544875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.544898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.544923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.544939 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.551781 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.568157 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.568223 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.568347 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.568250 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.568470 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.568599 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:35 crc kubenswrapper[4796]: E1127 11:25:35.568664 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.589743 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.606300 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.624943 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.642777 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.647854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.647915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.647933 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.647957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.647974 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.655923 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.691854 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:33Z\\\",\\\"message\\\":\\\"Sending *v1.Pod event handler 6 for removal\\\\nI1127 11:25:33.705752 6057 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 11:25:33.705779 6057 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 11:25:33.705796 6057 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 11:25:33.705810 6057 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 11:25:33.705819 6057 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 11:25:33.705421 6057 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:25:33.705483 6057 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:25:33.707013 6057 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 11:25:33.707049 6057 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 11:25:33.707071 6057 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:25:33.707158 6057 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:25:33.707983 6057 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.705817 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.729739 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:35Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.755429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.755473 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.755490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.755511 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.755528 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.859453 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.859539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.859562 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.859597 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.859622 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.962947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.963018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.963033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.963051 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:35 crc kubenswrapper[4796]: I1127 11:25:35.963063 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:35Z","lastTransitionTime":"2025-11-27T11:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.066175 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.066236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.066253 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.066306 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.066325 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.169203 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.169324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.169357 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.169392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.169410 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.272722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.272781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.272797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.272819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.272837 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.376387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.376485 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.376554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.376594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.376616 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.392439 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/1.log" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.393771 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/0.log" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.398365 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414" exitCode=1 Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.398442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.398526 4796 scope.go:117] "RemoveContainer" containerID="c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.400551 4796 scope.go:117] "RemoveContainer" containerID="e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414" Nov 27 11:25:36 crc kubenswrapper[4796]: E1127 11:25:36.400946 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.424314 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.445893 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.471464 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.480029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.480100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.480122 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.480146 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.480164 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.490687 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.512170 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.535209 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.547476 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.565726 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.568213 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:36 crc kubenswrapper[4796]: E1127 11:25:36.568379 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.583719 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.583793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.583819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.583848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.583873 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.584989 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.600000 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.612224 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.635451 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.652398 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.671254 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.686753 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.686786 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.686798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.686814 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.686825 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.690383 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.706459 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.737682 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1b9c82282d31327b1bc0de5848ef0cb355f9b1446bac02fab3f61420c32d52b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:33Z\\\",\\\"message\\\":\\\"Sending *v1.Pod event handler 6 for removal\\\\nI1127 11:25:33.705752 6057 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 11:25:33.705779 6057 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 11:25:33.705796 6057 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 11:25:33.705810 6057 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 11:25:33.705819 6057 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 11:25:33.705421 6057 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:25:33.705483 6057 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:25:33.707013 6057 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 11:25:33.707049 6057 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 11:25:33.707071 6057 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:25:33.707158 6057 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:25:33.707983 6057 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:36Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.789644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.789703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.789724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.789747 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.789765 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.893045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.893117 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.893134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.893158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.893176 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.996311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.996371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.996390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.996432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:36 crc kubenswrapper[4796]: I1127 11:25:36.996472 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:36Z","lastTransitionTime":"2025-11-27T11:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.099015 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.099081 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.099104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.099132 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.099154 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.202239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.202344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.202363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.202389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.202411 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.305572 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.305632 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.305651 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.305677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.305695 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.403747 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/1.log" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.407698 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.407756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.407779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.407809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.407832 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.408611 4796 scope.go:117] "RemoveContainer" containerID="e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414" Nov 27 11:25:37 crc kubenswrapper[4796]: E1127 11:25:37.408783 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.425556 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.441971 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.455415 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.472196 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.488043 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.512020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.512082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.512105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.512137 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.512159 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.521762 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.539196 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.569614 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.569609 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.569801 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:37 crc kubenswrapper[4796]: E1127 11:25:37.569930 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:37 crc kubenswrapper[4796]: E1127 11:25:37.570239 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:37 crc kubenswrapper[4796]: E1127 11:25:37.570424 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.582907 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.610363 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.619592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.619661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.619682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.619709 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.619726 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.650508 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.676731 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.698040 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.712523 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.721799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.721861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.721884 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.721909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.721928 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.725593 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.735336 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.749978 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.762792 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:37Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.824894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.824982 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.825002 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.825029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.825047 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.928539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.928606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.928626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.928652 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:37 crc kubenswrapper[4796]: I1127 11:25:37.928669 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:37Z","lastTransitionTime":"2025-11-27T11:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.031658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.031726 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.031751 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.031780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.031804 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.134445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.134517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.134541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.134563 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.134580 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.237825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.237883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.237901 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.237927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.237945 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.341091 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.341151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.341171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.341198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.341219 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.444101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.444171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.444197 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.444225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.444244 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.546317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.546372 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.546389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.546416 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.546434 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.568980 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:38 crc kubenswrapper[4796]: E1127 11:25:38.569148 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.649854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.649913 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.649930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.649953 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.649971 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.671476 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.687600 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.700954 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.721248 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.741745 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.752634 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.752677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.752694 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.752716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.752738 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.777026 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.801701 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.822052 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.844807 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.855237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.855318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.855336 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.855359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.855377 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.863256 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.890226 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.905705 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.927713 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.949411 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.958106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.958178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.958199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.958229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.958250 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:38Z","lastTransitionTime":"2025-11-27T11:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.970405 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:38 crc kubenswrapper[4796]: I1127 11:25:38.986231 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:38Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.005037 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:39Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.024195 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:39Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.044016 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:39Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.061320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.061363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.061373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.061389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.061401 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.164261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.164364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.164404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.164432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.164452 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.267064 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.267114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.267130 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.267150 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.267168 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.370432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.370494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.370512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.370537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.370555 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.473431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.473490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.473507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.473530 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.473549 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.569166 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:39 crc kubenswrapper[4796]: E1127 11:25:39.569409 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.569540 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.569674 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:39 crc kubenswrapper[4796]: E1127 11:25:39.569690 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:39 crc kubenswrapper[4796]: E1127 11:25:39.569825 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.576681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.576716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.576728 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.576743 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.576755 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.679482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.679540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.679557 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.679582 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.679600 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.782672 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.782712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.782722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.782737 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.782747 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.885667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.885737 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.885754 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.885777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.885799 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.989018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.989087 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.989103 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.989126 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:39 crc kubenswrapper[4796]: I1127 11:25:39.989142 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:39Z","lastTransitionTime":"2025-11-27T11:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.092539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.092578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.092589 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.092606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.092621 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.195938 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.195984 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.195994 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.196009 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.196024 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.299364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.299423 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.299441 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.299464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.299481 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.402015 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.402065 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.402082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.402107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.402124 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.505380 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.505426 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.505437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.505454 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.505465 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.515060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.515101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.515112 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.515127 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.515138 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.528543 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:40Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.533381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.533428 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.533436 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.533449 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.533458 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.547261 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:40Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.552330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.552381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.552393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.552413 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.552427 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.568332 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.568511 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.574317 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:40Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.578968 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.579018 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.579036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.579057 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.579076 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.601763 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:40Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.607556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.607597 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.607609 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.607627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.607640 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.622807 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:40Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:40 crc kubenswrapper[4796]: E1127 11:25:40.623075 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.625294 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.625331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.625343 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.625360 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.625374 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.729703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.729756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.729772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.729795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.729814 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.833678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.833788 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.833807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.833874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.833892 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.937392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.937441 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.937451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.937469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:40 crc kubenswrapper[4796]: I1127 11:25:40.937482 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:40Z","lastTransitionTime":"2025-11-27T11:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.040950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.041049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.041067 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.041125 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.041142 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.144802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.144866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.144883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.144907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.144925 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.248562 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.248650 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.248670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.248705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.248726 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.352602 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.352663 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.352681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.352706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.352723 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.456085 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.456160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.456180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.456209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.456236 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.560191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.560261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.560320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.560352 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.560375 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.568531 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.568582 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:41 crc kubenswrapper[4796]: E1127 11:25:41.568759 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.568820 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:41 crc kubenswrapper[4796]: E1127 11:25:41.568966 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:41 crc kubenswrapper[4796]: E1127 11:25:41.569133 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.663236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.663349 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.663367 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.663765 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.663819 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.767784 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.767834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.767853 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.767874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.767893 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.871152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.871215 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.871240 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.871319 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.871339 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.974448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.974539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.974556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.974579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:41 crc kubenswrapper[4796]: I1127 11:25:41.974598 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:41Z","lastTransitionTime":"2025-11-27T11:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.078362 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.078429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.078451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.078482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.078505 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.182001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.182054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.182071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.182092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.182109 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.284783 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.284827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.284840 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.284858 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.284871 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.387878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.387979 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.388004 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.388038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.388061 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.490986 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.491064 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.491092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.491121 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.491141 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.568372 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:42 crc kubenswrapper[4796]: E1127 11:25:42.568530 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.594839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.594910 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.594932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.594961 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.594983 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.698089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.698154 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.698172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.698199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.698219 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.801026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.801088 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.801110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.801180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.801206 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.904003 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.904061 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.904078 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.904100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:42 crc kubenswrapper[4796]: I1127 11:25:42.904117 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:42Z","lastTransitionTime":"2025-11-27T11:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.007241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.007280 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.007288 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.007300 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.007310 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.111949 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.112063 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.112136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.112243 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.112324 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.216114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.216180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.216201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.216229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.216249 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.318847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.318888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.318900 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.318926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.318939 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.421291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.421341 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.421358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.421390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.421409 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.523493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.523780 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.523789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.523802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.523811 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.568094 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.568113 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.568225 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:43 crc kubenswrapper[4796]: E1127 11:25:43.568385 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:43 crc kubenswrapper[4796]: E1127 11:25:43.568739 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:43 crc kubenswrapper[4796]: E1127 11:25:43.568990 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.599116 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.616822 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.626318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.626363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.626376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.626394 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.626410 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.633854 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.651615 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.667880 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.690471 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.709565 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.724931 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.729705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.729752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.729770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.729793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.729810 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.745128 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.766983 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.784369 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.800850 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.820664 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.832404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.832436 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.832446 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.832461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.832470 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.839560 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.851760 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.871100 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.885247 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.899517 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:43Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.935244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.935341 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.935360 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.935385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:43 crc kubenswrapper[4796]: I1127 11:25:43.935404 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:43Z","lastTransitionTime":"2025-11-27T11:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.038666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.038710 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.038722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.038739 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.038750 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.141571 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.141819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.141836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.141863 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.141880 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.246636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.246686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.246695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.246715 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.246726 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.349299 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.349362 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.349385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.349412 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.349437 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.453194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.453300 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.453321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.453351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.453374 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.555977 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.556022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.556032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.556047 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.556062 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.568674 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:44 crc kubenswrapper[4796]: E1127 11:25:44.568823 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.659026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.659092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.659102 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.659116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.659125 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.761866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.761919 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.761936 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.761960 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.761978 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.864176 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.864363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.864387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.864408 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.864426 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.967713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.967776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.967793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.967815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:44 crc kubenswrapper[4796]: I1127 11:25:44.967833 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:44Z","lastTransitionTime":"2025-11-27T11:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.070575 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.070645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.070664 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.070688 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.070705 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.174629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.174670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.174682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.174697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.174712 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.277466 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.277549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.277572 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.277607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.277630 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.380093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.380169 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.380192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.380221 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.380242 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.483098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.483165 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.483183 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.483211 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.483229 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.568433 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.568467 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.568464 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:45 crc kubenswrapper[4796]: E1127 11:25:45.568645 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:45 crc kubenswrapper[4796]: E1127 11:25:45.568836 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:45 crc kubenswrapper[4796]: E1127 11:25:45.568946 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.586311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.586360 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.586374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.586390 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.586402 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.689596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.689848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.689937 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.690010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.690072 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.793116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.793191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.793217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.793241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.793259 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.896263 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.897142 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.897326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.897520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:45 crc kubenswrapper[4796]: I1127 11:25:45.897674 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:45Z","lastTransitionTime":"2025-11-27T11:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.000472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.000896 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.001034 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.001163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.001311 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.104585 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.104901 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.105026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.105149 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.105316 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.208337 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.208426 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.208443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.208469 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.208488 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.310758 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.310801 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.310812 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.310830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.310843 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.414791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.414852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.414869 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.414892 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.414912 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.518427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.518495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.518517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.518542 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.518565 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.568972 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:46 crc kubenswrapper[4796]: E1127 11:25:46.569223 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.621423 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.621494 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.621516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.621543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.621566 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.648077 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.662524 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.694867 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.713795 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.723864 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.723928 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.723946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.723969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.723988 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.745893 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.772347 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.789742 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.806925 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.822896 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.827212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.827327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.827351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.827383 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.827420 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.836148 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.850825 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.865993 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.883211 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.898025 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.911754 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.923037 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.930255 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.930304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.930316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.930331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.930341 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:46Z","lastTransitionTime":"2025-11-27T11:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.938654 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.949448 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:46 crc kubenswrapper[4796]: I1127 11:25:46.958833 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:46Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.032897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.032941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.032955 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.032970 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.032979 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.135848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.135899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.135916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.135992 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.136011 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.242893 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.242976 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.242996 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.243020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.243037 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.345288 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.345321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.345329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.345342 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.345352 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.447711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.447763 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.447775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.447793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.447810 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.549931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.549990 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.550007 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.550030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.550047 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.568196 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.568365 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.568430 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:47 crc kubenswrapper[4796]: E1127 11:25:47.568718 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:47 crc kubenswrapper[4796]: E1127 11:25:47.568534 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:47 crc kubenswrapper[4796]: E1127 11:25:47.568823 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.653210 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.653250 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.653259 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.653285 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.653295 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.755538 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.755578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.755588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.755604 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.755615 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.858770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.859005 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.859090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.859205 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.859323 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.962074 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.962113 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.962123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.962138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:47 crc kubenswrapper[4796]: I1127 11:25:47.962149 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:47Z","lastTransitionTime":"2025-11-27T11:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.064616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.064681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.064699 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.064722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.064740 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.167504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.167553 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.167570 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.167593 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.167613 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.270314 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.270357 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.270368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.270385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.270397 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.373011 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.373083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.373100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.373123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.373140 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.476098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.476134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.476144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.476157 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.476166 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.568984 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:48 crc kubenswrapper[4796]: E1127 11:25:48.569189 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.578769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.578830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.578847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.578871 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.578889 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.681963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.682032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.682049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.682078 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.682098 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.785180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.785242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.785306 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.785341 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.785364 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.887920 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.887988 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.888012 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.888038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.888060 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.991006 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.991084 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.991111 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.991139 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:48 crc kubenswrapper[4796]: I1127 11:25:48.991162 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:48Z","lastTransitionTime":"2025-11-27T11:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.093984 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.094025 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.094036 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.094052 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.094064 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.196436 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.196504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.196521 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.196544 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.196563 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.299187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.299244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.299260 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.299309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.299326 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.401823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.401896 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.401912 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.401936 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.401966 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.504435 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.504495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.504512 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.504537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.504556 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.568641 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:49 crc kubenswrapper[4796]: E1127 11:25:49.568783 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.568880 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.568664 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:49 crc kubenswrapper[4796]: E1127 11:25:49.569036 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:49 crc kubenswrapper[4796]: E1127 11:25:49.569141 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.606924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.606973 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.606991 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.607125 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.607146 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.709864 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.709913 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.709926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.709946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.709965 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.812596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.812658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.812679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.812707 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.812729 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.915183 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.915261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.915303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.915326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:49 crc kubenswrapper[4796]: I1127 11:25:49.915345 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:49Z","lastTransitionTime":"2025-11-27T11:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.018718 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.018764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.018774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.018791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.018805 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.121225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.121252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.121260 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.121288 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.121298 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.223507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.223534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.223541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.223552 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.223560 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.325290 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.325349 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.325361 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.325375 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.325386 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.427202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.427260 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.427300 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.427317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.427328 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.530323 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.530378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.530392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.530409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.530421 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.568105 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.568374 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.632566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.632643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.632682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.632703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.632715 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.666116 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.666187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.666201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.666219 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.666248 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.681930 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:50Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.685335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.685376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.685389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.685404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.685417 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.698850 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:50Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.702774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.702817 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.702830 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.702844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.702853 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.718171 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:50Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.721898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.721936 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.721945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.721957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.721968 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.737216 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:50Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.741476 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.741523 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.741535 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.741552 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.741567 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.756820 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:50Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:50 crc kubenswrapper[4796]: E1127 11:25:50.756973 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.758627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.758716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.758773 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.758796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.758814 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.861382 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.861435 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.861451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.861470 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.861482 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.963554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.963607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.963619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.963633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:50 crc kubenswrapper[4796]: I1127 11:25:50.963644 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:50Z","lastTransitionTime":"2025-11-27T11:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.066604 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.066663 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.066681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.066709 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.066727 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.170300 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.170346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.170356 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.170373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.170383 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.273797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.273843 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.273854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.273873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.273885 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.336050 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:51 crc kubenswrapper[4796]: E1127 11:25:51.336170 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:51 crc kubenswrapper[4796]: E1127 11:25:51.336222 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:26:23.336207871 +0000 UTC m=+100.854526789 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.376140 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.376176 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.376186 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.376200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.376209 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.478100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.478136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.478146 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.478161 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.478173 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.568674 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.568720 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.568764 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:51 crc kubenswrapper[4796]: E1127 11:25:51.569182 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.569436 4796 scope.go:117] "RemoveContainer" containerID="e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414" Nov 27 11:25:51 crc kubenswrapper[4796]: E1127 11:25:51.569422 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:51 crc kubenswrapper[4796]: E1127 11:25:51.569560 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.580885 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.580942 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.580967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.580995 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.581014 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.684736 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.685214 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.685283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.685311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.685364 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.787445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.787479 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.787495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.787516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.787533 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.889858 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.889909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.889924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.889944 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.889960 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.992172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.992200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.992210 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.992224 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:51 crc kubenswrapper[4796]: I1127 11:25:51.992233 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:51Z","lastTransitionTime":"2025-11-27T11:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.097024 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.097074 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.097085 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.097101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.097114 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.199743 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.200066 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.200185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.200331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.200447 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.303385 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.303782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.303940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.304060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.304171 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.407117 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.407144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.407152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.407163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.407172 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.463590 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/1.log" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.466998 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.467585 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.479584 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.491822 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.503930 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.509839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.510071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.510256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.510513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.510660 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.520688 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.535784 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.550307 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.564155 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.568504 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:52 crc kubenswrapper[4796]: E1127 11:25:52.568695 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.576920 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.593942 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.607023 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.613986 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.614046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.614070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.614098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.614121 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.621320 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.632943 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.658074 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.674371 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.686082 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.700785 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.712425 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.716701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.716807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.716831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.716862 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.716885 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.734589 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:52Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.819772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.819819 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.819837 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.819856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.819871 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.923185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.923240 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.923262 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.923342 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:52 crc kubenswrapper[4796]: I1127 11:25:52.923366 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:52Z","lastTransitionTime":"2025-11-27T11:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.026220 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.026257 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.026285 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.026301 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.026312 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.128941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.128973 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.128985 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.129000 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.129009 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.231514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.231551 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.231561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.231576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.231587 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.333991 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.334045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.334059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.334076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.334088 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.436461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.436509 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.436520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.436537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.436549 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.472704 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/2.log" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.473687 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/1.log" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.476401 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0" exitCode=1 Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.476455 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.476501 4796 scope.go:117] "RemoveContainer" containerID="e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.477782 4796 scope.go:117] "RemoveContainer" containerID="4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0" Nov 27 11:25:53 crc kubenswrapper[4796]: E1127 11:25:53.478070 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.493138 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.507055 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.521128 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.533768 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.539114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.539160 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.539171 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.539190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.539203 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.548405 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.562013 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.568800 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:53 crc kubenswrapper[4796]: E1127 11:25:53.568906 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.569279 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:53 crc kubenswrapper[4796]: E1127 11:25:53.569420 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.569489 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:53 crc kubenswrapper[4796]: E1127 11:25:53.569546 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.575810 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.586058 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.595439 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.612933 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.622420 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.639862 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.642514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.642554 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.642563 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.642578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.642589 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.657148 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.669338 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.686710 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.699559 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.711975 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.733608 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.744045 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.744879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.744908 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.744917 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.744930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.744938 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.766058 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.779248 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.792329 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.803205 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.813266 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.830061 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1dcc12e9533900a09492ea5954b9cfd33711e731199e350cd0091b1c5097414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:35Z\\\",\\\"message\\\":\\\"r 0 failed attempt(s)\\\\nI1127 11:25:35.543919 6344 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1127 11:25:35.543925 6344 services_controller.go:452] Built service openshift-cluster-version/cluster-version-operator per-node LB for network=default: []services.LB{}\\\\nI1127 11:25:35.542007 6344 ovnkube.go:599] Stopped ovnkube\\\\nI1127 11:25:35.542122 6344 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1127 11:25:35.542202 6344 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-zfrrg\\\\nI1127 11:25:35.544171 6344 services_controller.go:453] Built service openshift-cluster-version/cluster-version-operator template LB for network=default: []services.LB{}\\\\nI1127 11:25:35.544192 6344 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1127 11:25:35.544198 6344 services_controller.go:454] Service openshift-cluster-version/cluster-version-operator for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1127 11:25:35.544317 6344 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.841198 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.846922 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.846960 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.846974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.846991 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.847004 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.855039 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.872168 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.892799 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.910360 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.924570 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.941144 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.949467 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.949517 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.949530 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.949548 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.949561 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:53Z","lastTransitionTime":"2025-11-27T11:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.955765 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.971428 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.985394 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:53 crc kubenswrapper[4796]: I1127 11:25:53.997809 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:53Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.051501 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.051563 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.051580 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.051605 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.051623 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.154767 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.154800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.154808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.154822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.154831 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.257123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.257170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.257180 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.257196 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.257209 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.359431 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.359477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.359491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.359507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.359520 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.461759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.461799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.461810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.461825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.461839 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.481054 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/2.log" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.483985 4796 scope.go:117] "RemoveContainer" containerID="4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0" Nov 27 11:25:54 crc kubenswrapper[4796]: E1127 11:25:54.484134 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.494628 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.511528 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.522029 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.542671 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.555341 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.563307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.563337 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.563345 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.563358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.563369 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.568406 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:54 crc kubenswrapper[4796]: E1127 11:25:54.568671 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.570697 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.577927 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.582009 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.594382 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.608092 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.621677 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.638885 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.654569 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.665738 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.665778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.665789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.665803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.665816 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.666585 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.678931 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.689317 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.709135 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.719497 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.727544 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:54Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.767584 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.767626 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.767636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.767652 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.767664 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.870778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.870814 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.870825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.870841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.870852 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.972795 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.972838 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.972850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.972864 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:54 crc kubenswrapper[4796]: I1127 11:25:54.972875 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:54Z","lastTransitionTime":"2025-11-27T11:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.075353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.075411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.075425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.075445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.075457 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.178880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.178925 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.178938 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.178955 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.178966 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.280894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.280934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.280944 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.280957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.280967 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.383217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.383251 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.383261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.383294 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.383305 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.486084 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.486155 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.486178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.486206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.486228 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.488096 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/0.log" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.488144 4796 generic.go:334] "Generic (PLEG): container finished" podID="1f8355a9-d2c1-4c54-be9d-68ef66397560" containerID="7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663" exitCode=1 Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.488296 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerDied","Data":"7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.489128 4796 scope.go:117] "RemoveContainer" containerID="7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.502881 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.514557 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.526973 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.540943 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.557036 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"2025-11-27T11:25:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17\\\\n2025-11-27T11:25:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17 to /host/opt/cni/bin/\\\\n2025-11-27T11:25:10Z [verbose] multus-daemon started\\\\n2025-11-27T11:25:10Z [verbose] Readiness Indicator file check\\\\n2025-11-27T11:25:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.567421 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.568484 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.568505 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:55 crc kubenswrapper[4796]: E1127 11:25:55.568570 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.568622 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:55 crc kubenswrapper[4796]: E1127 11:25:55.568792 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:55 crc kubenswrapper[4796]: E1127 11:25:55.568913 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.580843 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.589016 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.589046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.589054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.589068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.589080 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.591749 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"771f6fee-01f3-4068-aec5-afea05be08d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.603473 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.618826 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.633412 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.642983 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.651956 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.661746 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.677766 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.689813 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.690741 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.690777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.690794 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.690810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.690822 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.706713 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.716652 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.728622 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:55Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.793081 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.793115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.793123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.793136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.793146 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.895242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.895283 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.895292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.895304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.895315 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.997752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.997791 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.997802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.997813 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:55 crc kubenswrapper[4796]: I1127 11:25:55.997821 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:55Z","lastTransitionTime":"2025-11-27T11:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.099749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.099787 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.099796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.099823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.099833 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.202324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.202396 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.202409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.202427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.202440 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.304738 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.304778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.304786 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.304800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.304811 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.407777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.407839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.407849 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.407862 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.407872 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.494100 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/0.log" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.494190 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerStarted","Data":"6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.509255 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.509325 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.509349 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.509369 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.509384 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.514250 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"2025-11-27T11:25:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17\\\\n2025-11-27T11:25:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17 to /host/opt/cni/bin/\\\\n2025-11-27T11:25:10Z [verbose] multus-daemon started\\\\n2025-11-27T11:25:10Z [verbose] Readiness Indicator file check\\\\n2025-11-27T11:25:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.533884 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.545742 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.560088 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.568861 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:56 crc kubenswrapper[4796]: E1127 11:25:56.569198 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.573987 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.587622 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.597433 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.607675 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.611201 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.611256 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.611305 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.611329 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.611349 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.616755 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"771f6fee-01f3-4068-aec5-afea05be08d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.628031 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.639693 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.649885 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.659888 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.668849 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.679143 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.702076 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.712081 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.713562 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.713615 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.713627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.713646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.713661 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.731420 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.741985 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:25:56Z is after 2025-08-24T17:21:41Z" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.815527 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.815569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.815582 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.815598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.815609 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.918041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.918090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.918105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.918124 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:56 crc kubenswrapper[4796]: I1127 11:25:56.918137 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:56Z","lastTransitionTime":"2025-11-27T11:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.020829 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.020867 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.020875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.020888 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.020899 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.124334 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.124383 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.124392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.124409 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.124419 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.228054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.228115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.228128 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.228145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.228161 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.330590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.330667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.330686 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.330711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.330730 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.433602 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.433643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.433656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.433672 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.433685 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.536678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.536730 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.536743 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.536764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.536779 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.568309 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.568374 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.568320 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:57 crc kubenswrapper[4796]: E1127 11:25:57.568495 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:57 crc kubenswrapper[4796]: E1127 11:25:57.568665 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:57 crc kubenswrapper[4796]: E1127 11:25:57.568843 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.639408 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.639463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.639480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.639503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.639520 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.742532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.742595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.742611 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.742633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.742649 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.844833 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.844882 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.844896 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.844914 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.844929 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.947238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.947370 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.947398 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.947424 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:57 crc kubenswrapper[4796]: I1127 11:25:57.947441 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:57Z","lastTransitionTime":"2025-11-27T11:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.049482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.049548 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.049565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.049588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.049606 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.152599 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.152667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.152690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.152718 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.152739 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.255825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.255904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.255922 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.255946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.255966 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.358239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.358307 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.358321 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.358340 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.358352 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.460408 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.460478 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.460503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.460534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.460557 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.563945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.564015 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.564034 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.564058 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.564077 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.568456 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:25:58 crc kubenswrapper[4796]: E1127 11:25:58.568637 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.666856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.666906 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.666916 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.666935 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.666946 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.770046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.770122 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.770144 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.770172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.770193 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.874095 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.874153 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.874169 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.874194 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.874211 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.977659 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.977713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.977729 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.977751 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:58 crc kubenswrapper[4796]: I1127 11:25:58.977768 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:58Z","lastTransitionTime":"2025-11-27T11:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.080462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.080526 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.080546 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.080572 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.080590 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.183753 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.183825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.183851 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.183880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.183901 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.287324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.287393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.287414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.287443 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.287464 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.391243 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.391364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.391389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.391418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.391440 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.494041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.494130 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.494156 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.494186 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.494208 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.568682 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.568716 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.568886 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:25:59 crc kubenswrapper[4796]: E1127 11:25:59.569030 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:25:59 crc kubenswrapper[4796]: E1127 11:25:59.569137 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:25:59 crc kubenswrapper[4796]: E1127 11:25:59.569334 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.596693 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.596759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.596782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.596808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.596831 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.699246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.699308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.699320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.699334 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.699346 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.802437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.802488 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.802504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.802528 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.802545 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.905339 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.905371 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.905379 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.905419 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:25:59 crc kubenswrapper[4796]: I1127 11:25:59.905428 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:25:59Z","lastTransitionTime":"2025-11-27T11:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.008632 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.008702 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.008728 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.008757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.008781 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.112418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.112508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.112534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.112565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.112590 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.215619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.215681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.215694 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.215706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.215715 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.318379 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.318421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.318437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.318458 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.318479 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.421690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.421763 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.421781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.421805 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.421826 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.524178 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.524251 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.524302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.524330 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.524354 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.568984 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:00 crc kubenswrapper[4796]: E1127 11:26:00.569392 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.627759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.627833 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.627855 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.627931 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.627959 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.730878 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.730923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.730934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.730950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.730961 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.833714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.833764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.833778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.833796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.833811 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.905965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.906028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.906047 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.906071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.906087 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: E1127 11:26:00.924358 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:00Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.929032 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.929114 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.929137 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.929172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.929196 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: E1127 11:26:00.949536 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:00Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.953508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.953568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.953581 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.953598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.953610 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:00 crc kubenswrapper[4796]: E1127 11:26:00.973463 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:00Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.977717 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.977815 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.977835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.977861 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:00 crc kubenswrapper[4796]: I1127 11:26:00.977880 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:00Z","lastTransitionTime":"2025-11-27T11:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: E1127 11:26:01.000478 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:00Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.005491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.005550 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.005569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.005596 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.005613 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: E1127 11:26:01.024831 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:01Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:01 crc kubenswrapper[4796]: E1127 11:26:01.025051 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.028400 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.028445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.028463 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.028489 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.028508 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.131644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.131769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.131839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.131875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.131898 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.234789 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.234837 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.234880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.234903 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.234921 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.338069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.338177 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.338238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.338319 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.338347 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.441457 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.441498 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.441507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.441522 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.441531 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.544660 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.544773 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.544804 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.544881 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.544908 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.569150 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.569203 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:01 crc kubenswrapper[4796]: E1127 11:26:01.569379 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.569400 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:01 crc kubenswrapper[4796]: E1127 11:26:01.569520 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:01 crc kubenswrapper[4796]: E1127 11:26:01.569669 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.648110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.648749 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.648769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.648792 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.648810 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.751419 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.751485 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.751507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.751538 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.751560 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.855136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.855236 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.855328 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.855411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.855435 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.958538 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.958587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.958603 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.958627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:01 crc kubenswrapper[4796]: I1127 11:26:01.958644 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:01Z","lastTransitionTime":"2025-11-27T11:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.060636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.060777 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.060798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.060824 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.060846 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.163663 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.163712 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.163729 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.163750 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.163766 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.266678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.266773 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.266839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.266863 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.266918 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.370420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.370479 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.370500 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.370527 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.370548 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.473985 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.474045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.474061 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.474085 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.474102 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.568942 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:02 crc kubenswrapper[4796]: E1127 11:26:02.569052 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.581057 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.581110 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.581127 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.581150 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.581242 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.684841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.684900 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.684919 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.684945 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.684963 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.787798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.787863 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.787881 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.787905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.787924 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.891225 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.891295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.891309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.891327 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.891377 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.995089 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.995158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.995198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.995229 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:02 crc kubenswrapper[4796]: I1127 11:26:02.995250 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:02Z","lastTransitionTime":"2025-11-27T11:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.098218 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.098263 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.098292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.098310 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.098322 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.201123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.201161 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.201172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.201189 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.201201 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.305003 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.305101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.305120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.305326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.305551 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.408381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.408424 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.408487 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.408513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.408529 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.511659 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.511713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.511733 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.511758 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.511776 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.568449 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.568537 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.568622 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:03 crc kubenswrapper[4796]: E1127 11:26:03.568837 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:03 crc kubenswrapper[4796]: E1127 11:26:03.568903 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:03 crc kubenswrapper[4796]: E1127 11:26:03.569001 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.590308 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.608689 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.614347 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.614418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.614441 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.614472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.614497 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.640317 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.656808 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.688715 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.710583 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.718163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.718249 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.718320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.718356 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.718416 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.729410 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.751176 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.767137 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.783011 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.797645 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.816999 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"2025-11-27T11:25:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17\\\\n2025-11-27T11:25:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17 to /host/opt/cni/bin/\\\\n2025-11-27T11:25:10Z [verbose] multus-daemon started\\\\n2025-11-27T11:25:10Z [verbose] Readiness Indicator file check\\\\n2025-11-27T11:25:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.821530 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.821560 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.821570 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.821584 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.821595 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.833788 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.854940 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.867516 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"771f6fee-01f3-4068-aec5-afea05be08d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.887583 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.908034 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.924258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.924328 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.924358 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.924378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.924393 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:03Z","lastTransitionTime":"2025-11-27T11:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.952947 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:03 crc kubenswrapper[4796]: I1127 11:26:03.981999 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:03Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.031162 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.031302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.031364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.031425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.031459 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.133706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.133755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.133767 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.133784 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.133796 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.236558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.236598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.236617 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.236638 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.236655 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.339209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.339320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.339344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.339375 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.339399 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.442317 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.442389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.442407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.442432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.442450 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.544628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.544678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.544699 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.544727 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.544748 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.568487 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:04 crc kubenswrapper[4796]: E1127 11:26:04.568690 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.647635 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.647740 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.647803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.647841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.647907 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.751575 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.751631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.751647 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.751670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.751687 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.854231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.854261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.854281 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.854294 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.854304 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.956756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.956814 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.956834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.956860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:04 crc kubenswrapper[4796]: I1127 11:26:04.956877 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:04Z","lastTransitionTime":"2025-11-27T11:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.059847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.059915 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.059932 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.059955 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.059973 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.163192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.163438 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.163481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.163515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.163541 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.266576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.266645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.266667 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.266695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.266717 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.369856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.369910 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.369926 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.369947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.369965 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.475212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.475278 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.475293 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.475312 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.475325 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.568568 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.568610 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.568685 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:05 crc kubenswrapper[4796]: E1127 11:26:05.568856 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:05 crc kubenswrapper[4796]: E1127 11:26:05.568949 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:05 crc kubenswrapper[4796]: E1127 11:26:05.569041 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.577497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.577574 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.577595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.577625 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.577646 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.680728 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.680764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.680776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.680794 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.680806 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.784020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.784082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.784092 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.784107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.784116 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.887796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.887875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.887899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.887930 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.887949 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.991456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.991525 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.991542 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.991566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:05 crc kubenswrapper[4796]: I1127 11:26:05.991584 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:05Z","lastTransitionTime":"2025-11-27T11:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.094942 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.095030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.095049 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.095082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.095105 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.199000 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.199068 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.199079 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.199101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.199115 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.302374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.302451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.302477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.302509 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.302532 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.406523 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.406588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.406606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.406631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.406647 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.509564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.509609 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.509627 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.509651 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.509667 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.568802 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:06 crc kubenswrapper[4796]: E1127 11:26:06.568968 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.612392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.612472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.612490 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.612514 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.612533 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.715119 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.715190 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.715202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.715230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.715252 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.817150 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.817192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.817204 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.817222 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.817233 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.919217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.919316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.919335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.919359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:06 crc kubenswrapper[4796]: I1127 11:26:06.919377 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:06Z","lastTransitionTime":"2025-11-27T11:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.021689 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.021747 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.021766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.021788 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.021804 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.124288 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.124349 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.124366 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.124391 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.124468 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.227019 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.227081 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.227098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.227482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.227537 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.330696 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.330774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.330797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.330827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.330847 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.433831 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.433899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.433917 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.433940 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.433958 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.521945 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.522137 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522184 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:11.522149511 +0000 UTC m=+149.040468469 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.522242 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522383 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522428 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.522442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522448 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522511 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522598 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:11.522538422 +0000 UTC m=+149.040857370 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522631 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:11.522618494 +0000 UTC m=+149.040937442 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.522965 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.523073 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:11.523049536 +0000 UTC m=+149.041368494 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.536434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.536495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.536515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.536538 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.536556 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.568744 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.568838 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.568777 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.568925 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.569063 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.569241 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.624162 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.624403 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.624445 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.624463 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:26:07 crc kubenswrapper[4796]: E1127 11:26:07.624535 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:11.624514742 +0000 UTC m=+149.142833690 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.638964 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.639079 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.639104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.639135 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.639157 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.741461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.741518 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.741536 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.741558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.741577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.844541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.844604 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.844621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.844644 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.844665 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.947421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.947492 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.947516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.947546 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:07 crc kubenswrapper[4796]: I1127 11:26:07.947572 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:07Z","lastTransitionTime":"2025-11-27T11:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.050716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.050779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.050796 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.050821 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.050843 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.154891 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.154957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.154978 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.155001 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.155020 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.257401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.257535 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.257556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.257579 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.257601 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.360869 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.360949 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.360969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.361010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.361145 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.464546 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.464610 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.464630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.464653 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.464669 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.567520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.567581 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.567607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.567635 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.567656 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.568403 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:08 crc kubenswrapper[4796]: E1127 11:26:08.568573 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.672214 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.672343 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.672374 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.672402 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.672422 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.775198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.775297 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.775324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.775353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.775375 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.878756 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.878822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.878840 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.878868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.878888 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.982401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.982468 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.982486 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.982510 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:08 crc kubenswrapper[4796]: I1127 11:26:08.982528 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:08Z","lastTransitionTime":"2025-11-27T11:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.085378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.085420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.085428 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.085441 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.085450 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.188859 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.188919 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.188938 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.188960 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.188977 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.291465 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.291531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.291550 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.291577 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.291595 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.394771 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.394841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.394860 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.394885 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.394905 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.498172 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.498231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.498243 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.498262 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.498303 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.569084 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:09 crc kubenswrapper[4796]: E1127 11:26:09.569300 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.569594 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:09 crc kubenswrapper[4796]: E1127 11:26:09.569742 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.569607 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:09 crc kubenswrapper[4796]: E1127 11:26:09.569869 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.570125 4796 scope.go:117] "RemoveContainer" containerID="4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0" Nov 27 11:26:09 crc kubenswrapper[4796]: E1127 11:26:09.570482 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.601779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.601833 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.601849 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.601870 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.601885 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.704630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.704706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.704727 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.704755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.704773 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.807892 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.807964 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.807984 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.808008 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.808026 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.911976 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.912060 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.912082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.912112 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:09 crc kubenswrapper[4796]: I1127 11:26:09.912140 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:09Z","lastTransitionTime":"2025-11-27T11:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.015450 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.015537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.015561 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.015590 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.015611 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.118656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.118715 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.118732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.118755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.118773 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.221907 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.221995 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.222019 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.222050 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.222071 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.325950 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.326023 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.326040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.326065 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.326084 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.429437 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.429516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.429539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.429565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.429588 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.532120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.532206 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.532223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.532258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.532313 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.568772 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:10 crc kubenswrapper[4796]: E1127 11:26:10.568964 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.636346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.636421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.636444 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.636472 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.636493 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.739703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.739782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.739809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.739845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.739865 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.843242 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.843335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.843355 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.843381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.843401 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.946839 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.946927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.946952 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.946977 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:10 crc kubenswrapper[4796]: I1127 11:26:10.946993 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:10Z","lastTransitionTime":"2025-11-27T11:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.050744 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.050825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.050848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.050874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.050894 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.161406 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.161479 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.161499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.161526 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.161546 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.183598 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.183671 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.183695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.183724 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.183746 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.205792 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.210587 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.210630 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.210649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.210674 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.210692 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.231314 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.236799 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.236854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.236873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.236899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.236917 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.257396 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.262883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.262936 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.262953 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.262976 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.262993 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.286157 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.291700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.291755 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.291775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.291798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.291818 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.311891 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:11Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.312113 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.314076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.314131 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.314149 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.314169 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.314184 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.417231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.417363 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.417389 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.417418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.417437 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.520356 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.520406 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.520425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.520447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.520463 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.568290 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.568371 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.568475 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.568648 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.568847 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:11 crc kubenswrapper[4796]: E1127 11:26:11.569016 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.624376 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.624447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.624464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.624491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.624511 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.727732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.727835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.727859 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.727909 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.727936 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.831810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.831881 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.831899 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.831928 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.831946 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.935459 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.935535 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.935553 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.935578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:11 crc kubenswrapper[4796]: I1127 11:26:11.935604 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:11Z","lastTransitionTime":"2025-11-27T11:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.038610 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.038663 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.038682 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.038704 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.038722 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.142643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.142698 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.142714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.142737 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.142754 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.246033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.246131 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.246155 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.246184 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.246205 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.349238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.349335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.349359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.349382 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.349399 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.452135 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.452191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.452208 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.452228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.452245 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.554721 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.554784 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.554806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.554835 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.554856 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.568497 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:12 crc kubenswrapper[4796]: E1127 11:26:12.568706 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.658159 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.658230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.658254 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.658324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.658347 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.761421 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.761489 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.761511 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.761539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.761556 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.865445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.865496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.865513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.865536 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.865554 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.967879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.967954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.967974 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.967997 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:12 crc kubenswrapper[4796]: I1127 11:26:12.968016 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:12Z","lastTransitionTime":"2025-11-27T11:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.070883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.070943 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.070961 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.070988 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.071012 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.174337 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.174377 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.174388 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.174406 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.174417 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.277670 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.277734 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.277750 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.277774 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.277793 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.381211 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.381298 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.381318 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.381343 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.381360 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.484452 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.484519 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.484540 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.484564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.484581 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.568224 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.568314 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:13 crc kubenswrapper[4796]: E1127 11:26:13.568447 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.568543 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:13 crc kubenswrapper[4796]: E1127 11:26:13.568751 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:13 crc kubenswrapper[4796]: E1127 11:26:13.568862 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.587480 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.588233 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.588486 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.588643 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.589384 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.589567 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.605670 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.623082 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.645592 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"2025-11-27T11:25:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17\\\\n2025-11-27T11:25:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17 to /host/opt/cni/bin/\\\\n2025-11-27T11:25:10Z [verbose] multus-daemon started\\\\n2025-11-27T11:25:10Z [verbose] Readiness Indicator file check\\\\n2025-11-27T11:25:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.664606 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.682706 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.692941 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.692985 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.692998 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.693019 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.693034 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.712041 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"771f6fee-01f3-4068-aec5-afea05be08d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.731252 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.746662 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.773893 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.792134 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.797367 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.797418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.797471 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.797496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.797513 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.807501 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.821793 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.849110 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.866596 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.890585 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.899951 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.900006 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.900022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.900047 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.900064 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:13Z","lastTransitionTime":"2025-11-27T11:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.908616 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.925114 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:13 crc kubenswrapper[4796]: I1127 11:26:13.939952 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:13Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.002594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.002651 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.002668 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.002691 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.002711 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.105589 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.105655 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.105672 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.105695 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.105715 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.207905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.207984 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.208008 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.208039 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.208063 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.311531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.311595 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.311633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.311673 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.311696 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.415439 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.415508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.415529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.415558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.415579 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.519491 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.519556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.519573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.519594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.519618 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.568655 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:14 crc kubenswrapper[4796]: E1127 11:26:14.568841 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.622883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.622934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.622946 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.622965 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.622979 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.725404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.725468 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.725487 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.725508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.725524 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.828616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.828679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.828703 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.828729 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.828751 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.931481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.931551 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.931583 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.931613 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:14 crc kubenswrapper[4796]: I1127 11:26:14.931632 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:14Z","lastTransitionTime":"2025-11-27T11:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.034445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.034516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.034539 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.034566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.034589 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.137244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.137346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.137364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.137387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.137406 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.240968 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.241037 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.241059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.241088 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.241110 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.343650 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.343725 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.343748 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.343775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.343801 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.447217 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.447326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.447351 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.447379 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.447396 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.550848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.551297 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.551322 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.551344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.551361 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.568809 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.568920 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:15 crc kubenswrapper[4796]: E1127 11:26:15.569072 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:15 crc kubenswrapper[4796]: E1127 11:26:15.569297 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.569121 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:15 crc kubenswrapper[4796]: E1127 11:26:15.569832 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.654810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.654985 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.655022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.655056 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.655082 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.758246 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.758341 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.758359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.758383 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.758399 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.861354 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.861428 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.861453 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.861480 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.861506 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.964339 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.964392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.964411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.964434 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:15 crc kubenswrapper[4796]: I1127 11:26:15.964450 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:15Z","lastTransitionTime":"2025-11-27T11:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.067593 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.067653 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.067671 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.067693 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.067710 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.170772 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.170822 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.170834 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.170852 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.170864 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.273425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.273497 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.273507 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.273520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.273529 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.377185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.377245 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.377263 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.377316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.377335 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.479947 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.480012 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.480029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.480054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.480073 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.568428 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:16 crc kubenswrapper[4796]: E1127 11:26:16.568647 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.583090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.583131 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.583148 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.583168 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.583183 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.686692 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.686759 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.686776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.686802 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.686828 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.789193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.789239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.789250 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.789291 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.789323 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.892731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.892797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.892823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.892854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.892878 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.995496 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.995547 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.995566 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.995592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:16 crc kubenswrapper[4796]: I1127 11:26:16.995608 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:16Z","lastTransitionTime":"2025-11-27T11:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.098716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.098783 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.098800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.098823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.098840 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.202158 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.202228 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.202250 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.202324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.202353 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.305513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.305585 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.305607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.305629 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.305646 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.408407 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.408466 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.408481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.408503 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.408519 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.510798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.510856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.510874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.510895 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.510912 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.568578 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.568606 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.568642 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:17 crc kubenswrapper[4796]: E1127 11:26:17.568965 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:17 crc kubenswrapper[4796]: E1127 11:26:17.569300 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:17 crc kubenswrapper[4796]: E1127 11:26:17.569352 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.613074 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.613147 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.613173 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.613237 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.613263 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.715875 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.715934 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.715951 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.715976 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.715994 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.819403 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.819468 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.819484 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.819508 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.819525 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.922924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.922985 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.923002 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.923024 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:17 crc kubenswrapper[4796]: I1127 11:26:17.923042 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:17Z","lastTransitionTime":"2025-11-27T11:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.025979 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.026058 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.026082 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.026115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.026138 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.130200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.130309 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.130331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.130356 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.130375 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.233304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.233406 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.233425 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.233456 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.233486 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.337316 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.337401 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.337423 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.337453 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.337474 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.440576 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.440656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.440675 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.440705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.440724 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.543856 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.543911 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.543929 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.543958 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.543975 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.568133 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:18 crc kubenswrapper[4796]: E1127 11:26:18.568340 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.646767 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.646825 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.646841 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.646905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.646923 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.750305 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.750372 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.750395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.750429 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.750457 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.855664 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.855739 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.855757 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.855782 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.855800 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.958994 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.959059 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.959076 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.959099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:18 crc kubenswrapper[4796]: I1127 11:26:18.959118 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:18Z","lastTransitionTime":"2025-11-27T11:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.061736 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.061807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.061829 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.061854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.061871 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.165230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.165347 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.165364 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.165387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.165404 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.268969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.269093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.269156 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.269192 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.269246 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.372648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.372718 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.372745 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.372776 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.372799 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.476352 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.476411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.476427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.476445 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.476457 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.568734 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:19 crc kubenswrapper[4796]: E1127 11:26:19.568940 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.569047 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:19 crc kubenswrapper[4796]: E1127 11:26:19.569203 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.568734 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:19 crc kubenswrapper[4796]: E1127 11:26:19.569421 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.579252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.579320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.579335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.579353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.579365 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.682588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.682641 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.682657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.682679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.682696 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.787447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.787620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.787653 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.787741 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.787815 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.891393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.891442 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.891451 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.891467 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.891476 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.994198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.994231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.994239 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.994251 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:19 crc kubenswrapper[4796]: I1127 11:26:19.994260 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:19Z","lastTransitionTime":"2025-11-27T11:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.097531 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.097607 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.097631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.097661 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.097688 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.200641 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.200707 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.200736 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.200766 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.200787 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.303368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.303433 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.303455 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.303481 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.303500 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.406927 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.407007 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.407029 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.407058 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.407080 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.510632 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.510671 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.510681 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.510699 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.510711 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.568530 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:20 crc kubenswrapper[4796]: E1127 11:26:20.568711 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.613700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.613807 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.613828 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.613887 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.613907 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.717549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.717623 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.717648 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.717700 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.717724 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.821129 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.821187 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.821226 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.821249 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.821311 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.924764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.924810 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.924827 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.924848 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:20 crc kubenswrapper[4796]: I1127 11:26:20.924867 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:20Z","lastTransitionTime":"2025-11-27T11:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.028959 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.029026 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.029043 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.029070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.029088 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.132199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.132314 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.132350 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.132378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.132400 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.235191 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.235258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.235302 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.235325 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.235342 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.338078 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.338169 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.338200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.338230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.338251 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.415038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.415098 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.415115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.415138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.415156 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.438120 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:21Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.443711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.443849 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.443874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.443898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.443950 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.472222 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:21Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.484326 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.484417 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.484438 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.484513 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.484533 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.506082 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:21Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.511666 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.511716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.511731 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.511750 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.511765 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.531848 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:21Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.538418 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.538495 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.538520 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.538551 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.538577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.559973 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T11:26:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e2096ad-1d72-4cc4-93b3-695942a724e4\\\",\\\"systemUUID\\\":\\\"71d23547-b1b3-4287-9b6a-deece17c4b3f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:21Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.560390 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.562770 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.562858 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.562874 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.562892 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.562906 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.568303 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.568338 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.568844 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.569029 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.569449 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.569526 4796 scope.go:117] "RemoveContainer" containerID="4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0" Nov 27 11:26:21 crc kubenswrapper[4796]: E1127 11:26:21.569760 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.665094 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.665384 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.665396 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.665411 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.665423 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.768790 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.768829 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.768847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.768873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.768890 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.871989 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.872025 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.872033 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.872046 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.872054 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.980828 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.980873 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.980883 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.980898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:21 crc kubenswrapper[4796]: I1127 11:26:21.980909 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:21Z","lastTransitionTime":"2025-11-27T11:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.082821 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.082868 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.082879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.082897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.082911 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.185200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.185252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.185299 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.185324 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.185342 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.288042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.288107 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.288120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.288138 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.288155 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.390493 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.390558 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.390575 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.390599 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.390614 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.493112 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.493179 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.493196 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.493227 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.493244 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.568672 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:22 crc kubenswrapper[4796]: E1127 11:26:22.569161 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.594510 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/3.log" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.595134 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.595182 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.595202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.595230 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.595250 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.596098 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/2.log" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.600097 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" exitCode=1 Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.600139 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.600172 4796 scope.go:117] "RemoveContainer" containerID="4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.601637 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:26:22 crc kubenswrapper[4796]: E1127 11:26:22.601977 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.621814 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.641223 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.658172 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"771f6fee-01f3-4068-aec5-afea05be08d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.678385 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.698957 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.699022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.699040 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.699065 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.699086 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.699567 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.718665 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.734709 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.748260 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.762087 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.790579 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:26:22Z\\\",\\\"message\\\":\\\".com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:26:22.511154 6885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1127 11:26:22.511198 6885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 11:26:22.511305 6885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 11:26:22.511331 6885 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 11:26:22.511341 6885 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 11:26:22.511085 6885 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:26:22.511349 6885 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 11:26:22.511761 6885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1127 11:26:22.511830 6885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1127 11:26:22.511840 6885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1127 11:26:22.511877 6885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 11:26:22.511879 6885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 11:26:22.511924 6885 factory.go:656] Stopping watch factory\\\\nI1127 11:26:22.512156 6885 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.808815 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.808922 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.808960 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.808989 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.809020 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.809036 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.843072 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.859455 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.876795 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.891441 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.901441 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.910626 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.911396 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.911424 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.911432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.911464 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.911476 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:22Z","lastTransitionTime":"2025-11-27T11:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.921386 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:22 crc kubenswrapper[4796]: I1127 11:26:22.937616 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"2025-11-27T11:25:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17\\\\n2025-11-27T11:25:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17 to /host/opt/cni/bin/\\\\n2025-11-27T11:25:10Z [verbose] multus-daemon started\\\\n2025-11-27T11:25:10Z [verbose] Readiness Indicator file check\\\\n2025-11-27T11:25:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:22Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.014227 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.014286 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.014295 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.014308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.014316 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.116966 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.117031 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.117048 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.117071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.117088 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.220395 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.220474 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.220504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.220532 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.220553 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.322711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.322779 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.322800 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.322826 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.322848 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.416219 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:23 crc kubenswrapper[4796]: E1127 11:26:23.416438 4796 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:26:23 crc kubenswrapper[4796]: E1127 11:26:23.416546 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs podName:0494bee3-7923-49de-8c4f-e0fa4ffad936 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.416521149 +0000 UTC m=+164.934840137 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs") pod "network-metrics-daemon-79ll4" (UID: "0494bee3-7923-49de-8c4f-e0fa4ffad936") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.425344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.425397 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.425417 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.425439 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.425457 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.527241 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.527340 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.527365 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.527392 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.527411 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.568666 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.568740 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.568666 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:23 crc kubenswrapper[4796]: E1127 11:26:23.568894 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:23 crc kubenswrapper[4796]: E1127 11:26:23.569115 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:23 crc kubenswrapper[4796]: E1127 11:26:23.569227 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.583668 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6121b8f-0abf-4f33-aedf-072bc1e4edfa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfc4d2ba48867571aff1354e83d52c0cc697702b3931db25b91a2d13e7a64c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1acd1a69c1bcefcef792037445fc17bea6f9dc0fa969f00dd8ceaa35a4b88d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d624c2f390ae50976362e9b63ffa816efdffae26ebee32e3b8c5cd3cdf4385c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a28d148cbc48f00a8540d568dd6313a3a9a8786d35010e5c385906155ab18a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.595359 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e10446bec20f89128893709d3623fa9c28d91ac1297cad88696daa5c62cf49bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.611659 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/3.log" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.616722 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8drhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f8355a9-d2c1-4c54-be9d-68ef66397560\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:55Z\\\",\\\"message\\\":\\\"2025-11-27T11:25:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17\\\\n2025-11-27T11:25:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_62991e3b-c409-421b-897a-76291f99bb17 to /host/opt/cni/bin/\\\\n2025-11-27T11:25:10Z [verbose] multus-daemon started\\\\n2025-11-27T11:25:10Z [verbose] Readiness Indicator file check\\\\n2025-11-27T11:25:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p9mjq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8drhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.630152 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.630258 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.630342 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.630372 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.630458 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.631854 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa09a212-197e-4df8-9a5f-54321531252b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59397a3df9e88b52094f2cd1f9e153af47551754e00ff86b661bd2438b0985fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afaf2af38f99abafd2a23677b6c280fb8aa1d92d3376d08a742920e793bdfd16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33af5f8acc4285ecc086aa7b787a7ffdbe5f7618784673ddaceb8ddc24f1d0e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f047372a9c69e810c14de204217aee2cc31aa1df2cf2210985231ec890324f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc06f2ad2dca770d168e3fc3aeafcbee950bd22a3ca9615c7d2f2562fea39c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://092425d67a389fe8dfb596bcc53032a4a15cc04f5d19c661f58210089b51c594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bbc45c164da7e9a84bc3a7de08e7e05419253db8b2130b007ede9f78ecb57186\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pnk4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zfrrg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.645202 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e606fa06-e313-4bb9-b2cc-84ff65829b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8f23abee7894301692feb9925af46520d8c143275bec9232f87cf2426cbb934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fjx2d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v9kkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.656833 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b79ebf94-5367-49bd-8927-32b06a0c9490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6572f0802005f9076827994f157c0ef8bd704c7cba1a4b66359c52bd899ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b196b9ff0cc5915c058df1c351acf1185a77306e7dbea1b66dbd00fb23590f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d0442c220e7c44edb8a4488cfb1a887ea348874022730d893b003da783824ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.667713 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"771f6fee-01f3-4068-aec5-afea05be08d2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20fa698619b3374c6f70c9f05b6ddcd3677e95062159957d07a86f662d828f39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ed82d76ab25c13fb57d5f4018eebb44996c56f94972a7bd5b73f391054de37e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.679535 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.690335 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-79ll4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0494bee3-7923-49de-8c4f-e0fa4ffad936\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcv6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-79ll4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.705925 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"le observer\\\\nW1127 11:25:03.447774 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1127 11:25:03.447977 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 11:25:03.449185 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1665401920/tls.crt::/tmp/serving-cert-1665401920/tls.key\\\\\\\"\\\\nI1127 11:25:03.742469 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1127 11:25:03.746986 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1127 11:25:03.747019 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1127 11:25:03.747044 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1127 11:25:03.747052 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1127 11:25:03.756259 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1127 11:25:03.756314 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756320 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1127 11:25:03.756325 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1127 11:25:03.756329 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1127 11:25:03.756333 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1127 11:25:03.756337 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1127 11:25:03.756639 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1127 11:25:03.758789 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.718500 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.730449 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88565635-1352-4c89-9a97-7cbe728d543e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd353832580b57e11763be515d42261715d19bc03ea7693b2e8bf7c924da87ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41aa2f5c7dcccc04f7826f1830a9cbfe09f5306aaf40d7a9ff93a476d61761da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsqhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jngs5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.733331 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.733381 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.733393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.733412 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.733425 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.752722 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9cef17d7-9643-44ec-ae40-cc46b875ed42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4534b616e300b6d1e8854e550f77dbde668b27a31270a39c4c8aa5bb12eac1d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:25:52Z\\\",\\\"message\\\":\\\"alse, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"4c1be812-05d3-4f45-91b5-a853a5c8de71\\\\\\\", Protocol:\\\\\\\"udp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_TCP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Swi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T11:26:22Z\\\",\\\"message\\\":\\\".com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1127 11:26:22.511154 6885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1127 11:26:22.511198 6885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 11:26:22.511305 6885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 11:26:22.511331 6885 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1127 11:26:22.511341 6885 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 11:26:22.511085 6885 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 11:26:22.511349 6885 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1127 11:26:22.511761 6885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1127 11:26:22.511830 6885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1127 11:26:22.511840 6885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1127 11:26:22.511877 6885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1127 11:26:22.511879 6885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 11:26:22.511924 6885 factory.go:656] Stopping watch factory\\\\nI1127 11:26:22.512156 6885 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T11:26:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgvrx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6vnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.766369 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4z4td" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e236427-8ad0-41e6-861c-ed26c11c80fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce0994258a599f25ef9fdc5a59875cbad1cf83235d7214851bb76fb433a489e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5w9pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4z4td\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.785120 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c24774ac-7039-4932-9869-96602d236631\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c264466a72bf796ca402db0390b960048a2d04802e043339c39be2617aa28a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23b6edc43c8c56395c4ebe944482819744e2d6752ef50c9bbc7d08f3c915723b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1447a58f5ce5d12e0e190cb6eb16a5d7dc3004ebb91e09753ad9b4b80a465484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://676579c67b7661d25e4aeb886121e77c292b3b9e7841116e678049d49d12e279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8d85c54d4992bb95a38787231d7f3e7da8ad21f11943e1356cd416509070ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5f160fffd1d9c24a247f7b97672775ef7e81782b212881bc4d52cc2dfa654ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4679f02be8e4885a1d9d14db9b0d9d04bbf11b014c4e8df23d5758c98d0bcd20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:45Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4776a5acb9db179167abe2bfa13be769fc03b78957a0bfaf2380e2d35fbf5a17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T11:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T11:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:24:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.805286 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://912cb457b23c4d7710674cd2af5303b0a2b8bfe7e898adc975df7c6c410820f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.826946 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.835943 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.836034 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.836061 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.836095 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.836120 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.845047 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95a5ab6c03aa6abc66abf407317809276421500245f2cf360e1d4dc39b7dd418\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.860356 4796 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ns7ft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e576f87-15c8-4a31-b997-5dd3c973e51d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T11:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc32cce66608eb6c64ea0cec432800307491201147dfe1fd716d8500ab2d1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T11:25:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T11:25:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ns7ft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T11:26:23Z is after 2025-08-24T17:21:41Z" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.938960 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.939023 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.939042 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.939067 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:23 crc kubenswrapper[4796]: I1127 11:26:23.939084 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:23Z","lastTransitionTime":"2025-11-27T11:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.042222 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.042303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.042320 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.042343 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.042360 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.145373 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.145426 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.145448 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.145477 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.145499 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.248128 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.248216 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.248235 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.248285 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.248302 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.351393 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.351462 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.351482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.351504 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.351523 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.454568 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.454618 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.454631 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.454652 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.454664 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.557778 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.557845 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.557866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.557891 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.557912 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.568478 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:24 crc kubenswrapper[4796]: E1127 11:26:24.568698 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.661047 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.661093 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.661106 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.661123 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.661135 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.764805 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.764879 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.764897 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.764921 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.764938 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.867741 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.867787 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.867821 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.867836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.867848 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.970547 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.970584 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.970592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.970606 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:24 crc kubenswrapper[4796]: I1127 11:26:24.970614 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:24Z","lastTransitionTime":"2025-11-27T11:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.073705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.073764 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.073781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.073808 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.073834 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.177104 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.177173 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.177198 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.177231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.177253 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.280844 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.280902 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.280921 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.280944 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.280961 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.384223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.384338 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.384355 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.384382 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.384402 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.488145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.488213 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.488232 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.488299 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.488322 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.568343 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.568343 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:25 crc kubenswrapper[4796]: E1127 11:26:25.568497 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.568508 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:25 crc kubenswrapper[4796]: E1127 11:26:25.568776 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:25 crc kubenswrapper[4796]: E1127 11:26:25.568909 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.591529 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.591638 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.591657 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.591687 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.591705 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.695565 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.695642 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.695669 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.695701 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.695725 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.798722 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.798781 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.798797 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.798895 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.798916 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.902499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.902573 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.902592 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.902620 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:25 crc kubenswrapper[4796]: I1127 11:26:25.902637 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:25Z","lastTransitionTime":"2025-11-27T11:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.005646 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.005716 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.005735 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.005760 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.005779 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.107677 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.107735 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.107752 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.107775 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.107794 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.214209 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.214244 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.214253 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.214292 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.214310 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.317296 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.317346 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.317359 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.317378 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.317390 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.420120 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.420186 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.420204 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.420227 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.420244 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.523067 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.523141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.523163 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.523185 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.523202 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.569115 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:26 crc kubenswrapper[4796]: E1127 11:26:26.569313 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.624966 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.625041 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.625069 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.625100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.625122 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.727559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.727619 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.727636 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.727658 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.727676 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.831173 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.831233 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.831252 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.831311 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.831329 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.934030 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.934072 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.934087 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.934105 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:26 crc kubenswrapper[4796]: I1127 11:26:26.934118 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:26Z","lastTransitionTime":"2025-11-27T11:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.036741 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.036783 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.036793 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.036806 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.036816 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.139656 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.139687 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.139697 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.139711 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.139722 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.242482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.242547 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.242564 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.242588 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.242605 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.345361 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.345420 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.345432 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.345447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.345459 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.448212 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.448284 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.448298 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.448313 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.448323 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.550522 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.550569 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.550582 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.550599 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.550610 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.568069 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.568097 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:27 crc kubenswrapper[4796]: E1127 11:26:27.568201 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.568241 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:27 crc kubenswrapper[4796]: E1127 11:26:27.568396 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:27 crc kubenswrapper[4796]: E1127 11:26:27.568482 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.653549 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.653613 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.653628 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.653651 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.653669 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.763847 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.763904 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.763921 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.763944 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.763964 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.877115 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.877200 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.877238 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.877298 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.877331 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.980594 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.980678 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.980705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.980734 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:27 crc kubenswrapper[4796]: I1127 11:26:27.980757 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:27Z","lastTransitionTime":"2025-11-27T11:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.082938 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.082981 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.082992 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.083009 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.083020 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.185866 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.185913 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.185924 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.185943 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.185954 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.289145 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.289199 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.289213 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.289231 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.289243 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.392028 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.392079 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.392096 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.392117 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.392134 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.494608 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.494673 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.494690 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.494714 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.494738 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.568421 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:28 crc kubenswrapper[4796]: E1127 11:26:28.568637 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.597460 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.597521 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.597543 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.597570 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.597592 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.699836 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.699969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.699989 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.700010 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.700030 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.803452 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.803515 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.803534 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.803562 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.803580 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.906649 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.906706 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.906727 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.906751 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:28 crc kubenswrapper[4796]: I1127 11:26:28.906770 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:28Z","lastTransitionTime":"2025-11-27T11:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.068738 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.068846 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.068867 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.068891 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.068908 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.172732 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.172823 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.172838 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.172854 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.172894 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.276023 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.276071 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.276083 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.276129 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.276142 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.378733 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.378803 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.378814 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.378869 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.378887 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.483335 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.483387 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.483404 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.483427 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.483443 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.568223 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:29 crc kubenswrapper[4796]: E1127 11:26:29.568506 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.568349 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.568349 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:29 crc kubenswrapper[4796]: E1127 11:26:29.568628 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:29 crc kubenswrapper[4796]: E1127 11:26:29.568767 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.587141 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.587202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.587222 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.587247 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.587314 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.691132 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.691177 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.691188 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.691204 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.691216 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.794193 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.794249 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.794261 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.794303 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.794321 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.897447 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.897850 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.898006 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.898151 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:29 crc kubenswrapper[4796]: I1127 11:26:29.898307 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:29Z","lastTransitionTime":"2025-11-27T11:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.002625 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.002713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.002740 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.002768 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.002785 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.105967 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.106038 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.106061 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.106090 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.106111 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.213809 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.213880 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.213898 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.213923 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.213943 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.317250 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.318157 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.318499 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.318707 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.318854 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.421978 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.422054 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.422075 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.422099 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.422117 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.525482 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.525537 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.525553 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.525578 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.525596 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.568127 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:30 crc kubenswrapper[4796]: E1127 11:26:30.568412 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.627905 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.627969 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.627984 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.628005 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.628025 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.730954 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.731027 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.731045 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.731070 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.731088 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.834043 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.834096 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.834112 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.834136 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.834153 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.937541 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.937613 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.937638 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.937668 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:30 crc kubenswrapper[4796]: I1127 11:26:30.937692 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:30Z","lastTransitionTime":"2025-11-27T11:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.040633 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.040679 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.040694 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.040713 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.040738 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.143559 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.143616 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.143637 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.143664 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.143686 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.247624 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.247705 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.247769 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.247798 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.247821 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.360308 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.360344 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.360353 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.360368 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.360378 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.463100 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.463223 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.463304 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.463334 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.463351 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.566719 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.567022 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.567254 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.567516 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.567727 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.568052 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.568095 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.568110 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:31 crc kubenswrapper[4796]: E1127 11:26:31.568252 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:31 crc kubenswrapper[4796]: E1127 11:26:31.568663 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:31 crc kubenswrapper[4796]: E1127 11:26:31.568949 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.671894 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.672202 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.672483 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.672645 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.672786 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.776101 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.776461 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.776510 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.776556 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.776577 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.871414 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.871735 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.871963 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.872170 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.872447 4796 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T11:26:31Z","lastTransitionTime":"2025-11-27T11:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.940052 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q"] Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.940752 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.943573 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.944172 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.944434 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.944995 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.978845 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=84.978819057 podStartE2EDuration="1m24.978819057s" podCreationTimestamp="2025-11-27 11:25:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:31.967631199 +0000 UTC m=+109.485950157" watchObservedRunningTime="2025-11-27 11:26:31.978819057 +0000 UTC m=+109.497138015" Nov 27 11:26:31 crc kubenswrapper[4796]: I1127 11:26:31.993808 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=37.993788863 podStartE2EDuration="37.993788863s" podCreationTimestamp="2025-11-27 11:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:31.979298211 +0000 UTC m=+109.497617139" watchObservedRunningTime="2025-11-27 11:26:31.993788863 +0000 UTC m=+109.512107781" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.001866 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2a79436-d175-4dd5-b4bc-b000f7354087-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.002073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e2a79436-d175-4dd5-b4bc-b000f7354087-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.002118 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2a79436-d175-4dd5-b4bc-b000f7354087-service-ca\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.002140 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2a79436-d175-4dd5-b4bc-b000f7354087-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.002191 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e2a79436-d175-4dd5-b4bc-b000f7354087-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.028307 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.028242783 podStartE2EDuration="1m28.028242783s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.026416291 +0000 UTC m=+109.544735219" watchObservedRunningTime="2025-11-27 11:26:32.028242783 +0000 UTC m=+109.546561741" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.075260 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jngs5" podStartSLOduration=88.075237229 podStartE2EDuration="1m28.075237229s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.061102787 +0000 UTC m=+109.579421765" watchObservedRunningTime="2025-11-27 11:26:32.075237229 +0000 UTC m=+109.593556167" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.092185 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4z4td" podStartSLOduration=89.092164061 podStartE2EDuration="1m29.092164061s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.075816237 +0000 UTC m=+109.594135155" watchObservedRunningTime="2025-11-27 11:26:32.092164061 +0000 UTC m=+109.610482979" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.103499 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e2a79436-d175-4dd5-b4bc-b000f7354087-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.103559 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2a79436-d175-4dd5-b4bc-b000f7354087-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.103597 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e2a79436-d175-4dd5-b4bc-b000f7354087-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.103620 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2a79436-d175-4dd5-b4bc-b000f7354087-service-ca\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.103637 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2a79436-d175-4dd5-b4bc-b000f7354087-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.103909 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e2a79436-d175-4dd5-b4bc-b000f7354087-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.104140 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e2a79436-d175-4dd5-b4bc-b000f7354087-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.104830 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2a79436-d175-4dd5-b4bc-b000f7354087-service-ca\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.115409 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2a79436-d175-4dd5-b4bc-b000f7354087-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.119360 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2a79436-d175-4dd5-b4bc-b000f7354087-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-glf4q\" (UID: \"e2a79436-d175-4dd5-b4bc-b000f7354087\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.120554 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=86.120542879 podStartE2EDuration="1m26.120542879s" podCreationTimestamp="2025-11-27 11:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.120008283 +0000 UTC m=+109.638327201" watchObservedRunningTime="2025-11-27 11:26:32.120542879 +0000 UTC m=+109.638861797" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.195324 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ns7ft" podStartSLOduration=89.195301385 podStartE2EDuration="1m29.195301385s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.17437952 +0000 UTC m=+109.692698448" watchObservedRunningTime="2025-11-27 11:26:32.195301385 +0000 UTC m=+109.713620313" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.207720 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=54.207700598 podStartE2EDuration="54.207700598s" podCreationTimestamp="2025-11-27 11:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.207370689 +0000 UTC m=+109.725689627" watchObservedRunningTime="2025-11-27 11:26:32.207700598 +0000 UTC m=+109.726019516" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.253223 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8drhq" podStartSLOduration=89.253197712 podStartE2EDuration="1m29.253197712s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.231416012 +0000 UTC m=+109.749734950" watchObservedRunningTime="2025-11-27 11:26:32.253197712 +0000 UTC m=+109.771516640" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.253457 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zfrrg" podStartSLOduration=88.253448889 podStartE2EDuration="1m28.253448889s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.25313017 +0000 UTC m=+109.771449088" watchObservedRunningTime="2025-11-27 11:26:32.253448889 +0000 UTC m=+109.771767817" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.259094 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.266084 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podStartSLOduration=89.266055958 podStartE2EDuration="1m29.266055958s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:32.264433381 +0000 UTC m=+109.782752319" watchObservedRunningTime="2025-11-27 11:26:32.266055958 +0000 UTC m=+109.784374886" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.568754 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:32 crc kubenswrapper[4796]: E1127 11:26:32.569300 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.644537 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" event={"ID":"e2a79436-d175-4dd5-b4bc-b000f7354087","Type":"ContainerStarted","Data":"64e622da7a5245dbfc6968211b7eed75d7a6e364a34bc5bb7f15444437eec80d"} Nov 27 11:26:32 crc kubenswrapper[4796]: I1127 11:26:32.644584 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" event={"ID":"e2a79436-d175-4dd5-b4bc-b000f7354087","Type":"ContainerStarted","Data":"ca20e12e20deb5bcb5e2d11cc055bc05b096c7c9f3e411d32a84f072f47389a1"} Nov 27 11:26:33 crc kubenswrapper[4796]: I1127 11:26:33.568390 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:33 crc kubenswrapper[4796]: I1127 11:26:33.568468 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:33 crc kubenswrapper[4796]: I1127 11:26:33.568491 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:33 crc kubenswrapper[4796]: E1127 11:26:33.569427 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:33 crc kubenswrapper[4796]: E1127 11:26:33.569561 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:33 crc kubenswrapper[4796]: E1127 11:26:33.569647 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:33 crc kubenswrapper[4796]: I1127 11:26:33.666861 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-glf4q" podStartSLOduration=90.666831533 podStartE2EDuration="1m30.666831533s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:26:33.665374321 +0000 UTC m=+111.183693269" watchObservedRunningTime="2025-11-27 11:26:33.666831533 +0000 UTC m=+111.185150491" Nov 27 11:26:33 crc kubenswrapper[4796]: I1127 11:26:33.897195 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:26:33 crc kubenswrapper[4796]: I1127 11:26:33.934770 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:26:33 crc kubenswrapper[4796]: E1127 11:26:33.934934 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:26:34 crc kubenswrapper[4796]: I1127 11:26:34.568974 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:34 crc kubenswrapper[4796]: E1127 11:26:34.569202 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:35 crc kubenswrapper[4796]: I1127 11:26:35.568616 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:35 crc kubenswrapper[4796]: I1127 11:26:35.568667 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:35 crc kubenswrapper[4796]: I1127 11:26:35.568714 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:35 crc kubenswrapper[4796]: E1127 11:26:35.568837 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:35 crc kubenswrapper[4796]: E1127 11:26:35.568987 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:35 crc kubenswrapper[4796]: E1127 11:26:35.569245 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:36 crc kubenswrapper[4796]: I1127 11:26:36.568336 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:36 crc kubenswrapper[4796]: E1127 11:26:36.568461 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:37 crc kubenswrapper[4796]: I1127 11:26:37.568314 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:37 crc kubenswrapper[4796]: I1127 11:26:37.568395 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:37 crc kubenswrapper[4796]: I1127 11:26:37.568345 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:37 crc kubenswrapper[4796]: E1127 11:26:37.568522 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:37 crc kubenswrapper[4796]: E1127 11:26:37.568609 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:37 crc kubenswrapper[4796]: E1127 11:26:37.568695 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:38 crc kubenswrapper[4796]: I1127 11:26:38.568783 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:38 crc kubenswrapper[4796]: E1127 11:26:38.568927 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:39 crc kubenswrapper[4796]: I1127 11:26:39.568079 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:39 crc kubenswrapper[4796]: I1127 11:26:39.568149 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:39 crc kubenswrapper[4796]: E1127 11:26:39.568247 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:39 crc kubenswrapper[4796]: E1127 11:26:39.568388 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:39 crc kubenswrapper[4796]: I1127 11:26:39.568439 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:39 crc kubenswrapper[4796]: E1127 11:26:39.568521 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:40 crc kubenswrapper[4796]: I1127 11:26:40.568599 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:40 crc kubenswrapper[4796]: E1127 11:26:40.568815 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.568925 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.569103 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:41 crc kubenswrapper[4796]: E1127 11:26:41.569166 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.569297 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:41 crc kubenswrapper[4796]: E1127 11:26:41.569449 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:41 crc kubenswrapper[4796]: E1127 11:26:41.569545 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.673701 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/1.log" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.674361 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/0.log" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.674431 4796 generic.go:334] "Generic (PLEG): container finished" podID="1f8355a9-d2c1-4c54-be9d-68ef66397560" containerID="6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b" exitCode=1 Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.674511 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerDied","Data":"6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b"} Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.674632 4796 scope.go:117] "RemoveContainer" containerID="7c69670609d0db556886d1aa8ca4bde7ad9f6d12f31d3cf7f9358bbb5d867663" Nov 27 11:26:41 crc kubenswrapper[4796]: I1127 11:26:41.675176 4796 scope.go:117] "RemoveContainer" containerID="6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b" Nov 27 11:26:41 crc kubenswrapper[4796]: E1127 11:26:41.675426 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8drhq_openshift-multus(1f8355a9-d2c1-4c54-be9d-68ef66397560)\"" pod="openshift-multus/multus-8drhq" podUID="1f8355a9-d2c1-4c54-be9d-68ef66397560" Nov 27 11:26:42 crc kubenswrapper[4796]: I1127 11:26:42.568813 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:42 crc kubenswrapper[4796]: E1127 11:26:42.569087 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:42 crc kubenswrapper[4796]: I1127 11:26:42.678748 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/1.log" Nov 27 11:26:43 crc kubenswrapper[4796]: I1127 11:26:43.568952 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:43 crc kubenswrapper[4796]: I1127 11:26:43.568990 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:43 crc kubenswrapper[4796]: I1127 11:26:43.569095 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:43 crc kubenswrapper[4796]: E1127 11:26:43.570672 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:43 crc kubenswrapper[4796]: E1127 11:26:43.570794 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:43 crc kubenswrapper[4796]: E1127 11:26:43.570954 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:43 crc kubenswrapper[4796]: E1127 11:26:43.587325 4796 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 27 11:26:43 crc kubenswrapper[4796]: E1127 11:26:43.702132 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:26:44 crc kubenswrapper[4796]: I1127 11:26:44.568126 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:44 crc kubenswrapper[4796]: E1127 11:26:44.568339 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:45 crc kubenswrapper[4796]: I1127 11:26:45.568754 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:45 crc kubenswrapper[4796]: I1127 11:26:45.568793 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:45 crc kubenswrapper[4796]: E1127 11:26:45.568909 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:45 crc kubenswrapper[4796]: I1127 11:26:45.568979 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:45 crc kubenswrapper[4796]: E1127 11:26:45.569169 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:45 crc kubenswrapper[4796]: E1127 11:26:45.569313 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:46 crc kubenswrapper[4796]: I1127 11:26:46.568511 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:46 crc kubenswrapper[4796]: E1127 11:26:46.568671 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:47 crc kubenswrapper[4796]: I1127 11:26:47.568500 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:47 crc kubenswrapper[4796]: E1127 11:26:47.568632 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:47 crc kubenswrapper[4796]: I1127 11:26:47.568529 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:47 crc kubenswrapper[4796]: E1127 11:26:47.568823 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:47 crc kubenswrapper[4796]: I1127 11:26:47.568850 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:47 crc kubenswrapper[4796]: E1127 11:26:47.569672 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:47 crc kubenswrapper[4796]: I1127 11:26:47.569716 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:26:47 crc kubenswrapper[4796]: E1127 11:26:47.570085 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:26:48 crc kubenswrapper[4796]: I1127 11:26:48.568745 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:48 crc kubenswrapper[4796]: E1127 11:26:48.568947 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:48 crc kubenswrapper[4796]: E1127 11:26:48.703218 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:26:49 crc kubenswrapper[4796]: I1127 11:26:49.568497 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:49 crc kubenswrapper[4796]: I1127 11:26:49.568725 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:49 crc kubenswrapper[4796]: E1127 11:26:49.568742 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:49 crc kubenswrapper[4796]: E1127 11:26:49.568908 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:49 crc kubenswrapper[4796]: I1127 11:26:49.568998 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:49 crc kubenswrapper[4796]: E1127 11:26:49.569707 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:50 crc kubenswrapper[4796]: I1127 11:26:50.569025 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:50 crc kubenswrapper[4796]: E1127 11:26:50.569417 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:51 crc kubenswrapper[4796]: I1127 11:26:51.568516 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:51 crc kubenswrapper[4796]: I1127 11:26:51.568576 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:51 crc kubenswrapper[4796]: I1127 11:26:51.568638 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:51 crc kubenswrapper[4796]: E1127 11:26:51.570211 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:51 crc kubenswrapper[4796]: E1127 11:26:51.570544 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:51 crc kubenswrapper[4796]: E1127 11:26:51.570546 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:52 crc kubenswrapper[4796]: I1127 11:26:52.568604 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:52 crc kubenswrapper[4796]: E1127 11:26:52.568786 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:53 crc kubenswrapper[4796]: I1127 11:26:53.568427 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:53 crc kubenswrapper[4796]: I1127 11:26:53.568465 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:53 crc kubenswrapper[4796]: E1127 11:26:53.570111 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:53 crc kubenswrapper[4796]: I1127 11:26:53.570164 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:53 crc kubenswrapper[4796]: E1127 11:26:53.570430 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:53 crc kubenswrapper[4796]: E1127 11:26:53.570304 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:53 crc kubenswrapper[4796]: E1127 11:26:53.704007 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:26:54 crc kubenswrapper[4796]: I1127 11:26:54.568375 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:54 crc kubenswrapper[4796]: E1127 11:26:54.568548 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:55 crc kubenswrapper[4796]: I1127 11:26:55.569034 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:55 crc kubenswrapper[4796]: I1127 11:26:55.569072 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:55 crc kubenswrapper[4796]: E1127 11:26:55.569322 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:55 crc kubenswrapper[4796]: I1127 11:26:55.569514 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:55 crc kubenswrapper[4796]: E1127 11:26:55.570114 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:55 crc kubenswrapper[4796]: E1127 11:26:55.571032 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:56 crc kubenswrapper[4796]: I1127 11:26:56.569090 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:56 crc kubenswrapper[4796]: E1127 11:26:56.569576 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:56 crc kubenswrapper[4796]: I1127 11:26:56.569857 4796 scope.go:117] "RemoveContainer" containerID="6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b" Nov 27 11:26:57 crc kubenswrapper[4796]: I1127 11:26:57.568861 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:57 crc kubenswrapper[4796]: I1127 11:26:57.568909 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:57 crc kubenswrapper[4796]: I1127 11:26:57.569073 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:57 crc kubenswrapper[4796]: E1127 11:26:57.569320 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:57 crc kubenswrapper[4796]: E1127 11:26:57.569398 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:57 crc kubenswrapper[4796]: E1127 11:26:57.569447 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:57 crc kubenswrapper[4796]: I1127 11:26:57.736445 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/1.log" Nov 27 11:26:57 crc kubenswrapper[4796]: I1127 11:26:57.736543 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerStarted","Data":"12e4f74214e01b5188063377438e0ba34315407131a3a43c591387bc0018e91b"} Nov 27 11:26:58 crc kubenswrapper[4796]: I1127 11:26:58.568517 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:26:58 crc kubenswrapper[4796]: E1127 11:26:58.568624 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:26:58 crc kubenswrapper[4796]: E1127 11:26:58.706053 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:26:59 crc kubenswrapper[4796]: I1127 11:26:59.568827 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:26:59 crc kubenswrapper[4796]: E1127 11:26:59.569025 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:26:59 crc kubenswrapper[4796]: I1127 11:26:59.569339 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:26:59 crc kubenswrapper[4796]: E1127 11:26:59.569445 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:26:59 crc kubenswrapper[4796]: I1127 11:26:59.570221 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:26:59 crc kubenswrapper[4796]: E1127 11:26:59.570538 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:26:59 crc kubenswrapper[4796]: I1127 11:26:59.570853 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:26:59 crc kubenswrapper[4796]: E1127 11:26:59.571206 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6vnft_openshift-ovn-kubernetes(9cef17d7-9643-44ec-ae40-cc46b875ed42)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" Nov 27 11:27:00 crc kubenswrapper[4796]: I1127 11:27:00.568234 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:00 crc kubenswrapper[4796]: E1127 11:27:00.568451 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:01 crc kubenswrapper[4796]: I1127 11:27:01.568428 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:01 crc kubenswrapper[4796]: I1127 11:27:01.568492 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:01 crc kubenswrapper[4796]: E1127 11:27:01.568571 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:01 crc kubenswrapper[4796]: I1127 11:27:01.568677 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:01 crc kubenswrapper[4796]: E1127 11:27:01.568809 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:01 crc kubenswrapper[4796]: E1127 11:27:01.569036 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:02 crc kubenswrapper[4796]: I1127 11:27:02.568471 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:02 crc kubenswrapper[4796]: E1127 11:27:02.568623 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:03 crc kubenswrapper[4796]: I1127 11:27:03.569004 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:03 crc kubenswrapper[4796]: I1127 11:27:03.569071 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:03 crc kubenswrapper[4796]: I1127 11:27:03.569139 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:03 crc kubenswrapper[4796]: E1127 11:27:03.570096 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:03 crc kubenswrapper[4796]: E1127 11:27:03.570291 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:03 crc kubenswrapper[4796]: E1127 11:27:03.570329 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:03 crc kubenswrapper[4796]: E1127 11:27:03.706802 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:27:04 crc kubenswrapper[4796]: I1127 11:27:04.568557 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:04 crc kubenswrapper[4796]: E1127 11:27:04.568981 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:05 crc kubenswrapper[4796]: I1127 11:27:05.568620 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:05 crc kubenswrapper[4796]: I1127 11:27:05.568647 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:05 crc kubenswrapper[4796]: I1127 11:27:05.568693 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:05 crc kubenswrapper[4796]: E1127 11:27:05.569730 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:05 crc kubenswrapper[4796]: E1127 11:27:05.569810 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:05 crc kubenswrapper[4796]: E1127 11:27:05.569875 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:06 crc kubenswrapper[4796]: I1127 11:27:06.568444 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:06 crc kubenswrapper[4796]: E1127 11:27:06.568829 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:07 crc kubenswrapper[4796]: I1127 11:27:07.568813 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:07 crc kubenswrapper[4796]: I1127 11:27:07.568873 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:07 crc kubenswrapper[4796]: I1127 11:27:07.568813 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:07 crc kubenswrapper[4796]: E1127 11:27:07.568992 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:07 crc kubenswrapper[4796]: E1127 11:27:07.569081 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:07 crc kubenswrapper[4796]: E1127 11:27:07.569173 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:08 crc kubenswrapper[4796]: I1127 11:27:08.568898 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:08 crc kubenswrapper[4796]: E1127 11:27:08.569093 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:08 crc kubenswrapper[4796]: E1127 11:27:08.708346 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:27:09 crc kubenswrapper[4796]: I1127 11:27:09.569134 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:09 crc kubenswrapper[4796]: I1127 11:27:09.569134 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:09 crc kubenswrapper[4796]: E1127 11:27:09.569460 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:09 crc kubenswrapper[4796]: E1127 11:27:09.569564 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:09 crc kubenswrapper[4796]: I1127 11:27:09.570408 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:09 crc kubenswrapper[4796]: E1127 11:27:09.570754 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:10 crc kubenswrapper[4796]: I1127 11:27:10.569317 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:10 crc kubenswrapper[4796]: E1127 11:27:10.569572 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.556605 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.556808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.556887 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.556950 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:29:13.556918146 +0000 UTC m=+271.075237064 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557098 4796 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.557109 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557208 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:29:13.557175712 +0000 UTC m=+271.075494850 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557222 4796 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557282 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 11:29:13.557257234 +0000 UTC m=+271.075576402 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557390 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557402 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557414 4796 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.557439 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 11:29:13.557433339 +0000 UTC m=+271.075752257 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.568873 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.568914 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.569063 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.569085 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.569187 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.569303 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:11 crc kubenswrapper[4796]: I1127 11:27:11.658196 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.658481 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.658521 4796 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.658535 4796 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:27:11 crc kubenswrapper[4796]: E1127 11:27:11.658601 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 11:29:13.658582607 +0000 UTC m=+271.176901615 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 11:27:12 crc kubenswrapper[4796]: I1127 11:27:12.568828 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:12 crc kubenswrapper[4796]: E1127 11:27:12.569056 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:12 crc kubenswrapper[4796]: I1127 11:27:12.570105 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.568092 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.568096 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:13 crc kubenswrapper[4796]: E1127 11:27:13.569982 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.570023 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:13 crc kubenswrapper[4796]: E1127 11:27:13.570140 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:13 crc kubenswrapper[4796]: E1127 11:27:13.570251 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:13 crc kubenswrapper[4796]: E1127 11:27:13.709698 4796 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.795474 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/3.log" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.798851 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerStarted","Data":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.799370 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.844144 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podStartSLOduration=129.844114015 podStartE2EDuration="2m9.844114015s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:13.830854015 +0000 UTC m=+151.349172963" watchObservedRunningTime="2025-11-27 11:27:13.844114015 +0000 UTC m=+151.362432943" Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.845845 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-79ll4"] Nov 27 11:27:13 crc kubenswrapper[4796]: I1127 11:27:13.845952 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:13 crc kubenswrapper[4796]: E1127 11:27:13.846063 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:14 crc kubenswrapper[4796]: I1127 11:27:14.568679 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:14 crc kubenswrapper[4796]: E1127 11:27:14.568929 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:15 crc kubenswrapper[4796]: I1127 11:27:15.568495 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:15 crc kubenswrapper[4796]: I1127 11:27:15.568563 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:15 crc kubenswrapper[4796]: E1127 11:27:15.569085 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:15 crc kubenswrapper[4796]: E1127 11:27:15.568920 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:15 crc kubenswrapper[4796]: I1127 11:27:15.568607 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:15 crc kubenswrapper[4796]: E1127 11:27:15.569192 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:16 crc kubenswrapper[4796]: I1127 11:27:16.568515 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:16 crc kubenswrapper[4796]: E1127 11:27:16.568748 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:17 crc kubenswrapper[4796]: I1127 11:27:17.568751 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:17 crc kubenswrapper[4796]: I1127 11:27:17.568870 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:17 crc kubenswrapper[4796]: E1127 11:27:17.568942 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 11:27:17 crc kubenswrapper[4796]: I1127 11:27:17.568986 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:17 crc kubenswrapper[4796]: E1127 11:27:17.569120 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 11:27:17 crc kubenswrapper[4796]: E1127 11:27:17.569246 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-79ll4" podUID="0494bee3-7923-49de-8c4f-e0fa4ffad936" Nov 27 11:27:18 crc kubenswrapper[4796]: I1127 11:27:18.568922 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:18 crc kubenswrapper[4796]: E1127 11:27:18.569135 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.568706 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.568770 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.568730 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.570862 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.571443 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.571794 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 11:27:19 crc kubenswrapper[4796]: I1127 11:27:19.572170 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 11:27:20 crc kubenswrapper[4796]: I1127 11:27:20.568858 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:27:20 crc kubenswrapper[4796]: I1127 11:27:20.571050 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 11:27:20 crc kubenswrapper[4796]: I1127 11:27:20.572452 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.909621 4796 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.957298 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bxf5l"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.957884 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.957977 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.958434 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.958933 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.959250 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.959701 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.960100 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-skdgb"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.960830 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.960843 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.960858 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.961924 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.962787 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.962907 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.963245 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.963438 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rdzj8"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.964175 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.964541 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.964710 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.964911 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.964396 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.965240 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.965513 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.965551 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.965304 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.966175 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.966740 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.967962 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-bvccg"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.968658 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.969633 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971036 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971078 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971186 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971343 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971444 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971538 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971616 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971858 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971906 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971928 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.971985 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.972172 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.973488 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dp4q8"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.974238 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.974298 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.975654 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.986444 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zbnzg"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.986888 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.987347 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.987634 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.987759 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.995960 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.996207 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sngfr"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.996542 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.996846 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.996976 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997105 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997189 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997260 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997360 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997490 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997573 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.997875 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.998543 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.998634 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.998653 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.998767 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk"] Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.998984 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts2d7\" (UniqueName: \"kubernetes.io/projected/dbd3b91b-be30-474a-94d4-b4a684eef907-kube-api-access-ts2d7\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999009 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999028 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2mbh\" (UniqueName: \"kubernetes.io/projected/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-kube-api-access-l2mbh\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999044 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8096c4b-d1ba-471a-9d05-22aab449b51c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fgr9n\" (UID: \"f8096c4b-d1ba-471a-9d05-22aab449b51c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999068 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mnk4\" (UniqueName: \"kubernetes.io/projected/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-kube-api-access-9mnk4\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999091 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zq8t\" (UniqueName: \"kubernetes.io/projected/f9d82114-9bc6-4bce-8418-46c860542aa4-kube-api-access-7zq8t\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999106 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-serving-cert\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999121 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-encryption-config\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999137 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-service-ca-bundle\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999151 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999157 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-client-ca\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999172 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v762j\" (UniqueName: \"kubernetes.io/projected/d779a1a6-0749-48f9-a4a7-75e65a13c7da-kube-api-access-v762j\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999213 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dbd3b91b-be30-474a-94d4-b4a684eef907-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999228 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999243 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999259 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbd3b91b-be30-474a-94d4-b4a684eef907-serving-cert\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999295 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6s8q\" (UniqueName: \"kubernetes.io/projected/f8096c4b-d1ba-471a-9d05-22aab449b51c-kube-api-access-g6s8q\") pod \"cluster-samples-operator-665b6dd947-fgr9n\" (UID: \"f8096c4b-d1ba-471a-9d05-22aab449b51c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999309 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-config\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999324 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-etcd-client\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999339 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999355 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-config\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999371 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9d82114-9bc6-4bce-8418-46c860542aa4-audit-dir\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999398 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d779a1a6-0749-48f9-a4a7-75e65a13c7da-serving-cert\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999424 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp8rd\" (UniqueName: \"kubernetes.io/projected/ce3173af-2480-4a96-a7c5-ab9c44b18c64-kube-api-access-hp8rd\") pod \"downloads-7954f5f757-rdzj8\" (UID: \"ce3173af-2480-4a96-a7c5-ab9c44b18c64\") " pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999445 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-config\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999460 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-images\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999476 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-serving-cert\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999489 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d4pz\" (UniqueName: \"kubernetes.io/projected/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-kube-api-access-4d4pz\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999504 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-audit-policies\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999519 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:22 crc kubenswrapper[4796]: I1127 11:27:22.999822 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pmhfw"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.000223 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.000969 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l954k"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001372 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001440 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001641 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001770 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001832 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001871 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001956 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002031 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.001955 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002353 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002458 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002562 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002584 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002631 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002453 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002733 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.002758 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.006685 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.006866 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.007049 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.007189 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.007310 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.008293 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z8hd2"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.008733 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009063 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009172 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009252 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009603 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009682 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009757 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009854 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.009961 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.010049 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.010166 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.010599 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5brmj"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.010913 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.011125 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.011515 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.011581 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.012503 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.013134 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.013386 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.013637 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.013996 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.014403 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.014800 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.015246 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.016331 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.016627 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.018445 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fdwzj"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.018761 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-v65j9"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.018991 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.019082 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.019131 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.019649 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.019981 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.021224 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.023243 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.023547 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.025742 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.039982 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.040780 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.041641 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7rks7"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.056068 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.058816 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.058879 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.059512 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.059557 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.060070 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.060471 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.060625 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.060745 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.060849 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.060946 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061051 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061286 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061426 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061520 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061628 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061731 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061851 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.061937 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.062017 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.062813 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bxf5l"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.062850 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.068845 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.072447 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.073013 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ktgzm"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.073330 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.073349 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.073630 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-skdgb"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.073647 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.073952 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7t7jg"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.074628 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.074647 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.074657 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dp4q8"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.074737 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.074941 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.075095 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.075214 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.075414 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.075556 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.075872 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rdzj8"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.075953 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bvccg"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.076931 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-gqch4"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.077808 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.078478 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l954k"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.094752 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100370 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100412 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zq8t\" (UniqueName: \"kubernetes.io/projected/f9d82114-9bc6-4bce-8418-46c860542aa4-kube-api-access-7zq8t\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100449 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr7bs\" (UniqueName: \"kubernetes.io/projected/222d84fa-7732-4a9d-b250-e66c95092ad6-kube-api-access-lr7bs\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100472 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-serving-cert\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100487 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-oauth-config\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100520 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474fx\" (UniqueName: \"kubernetes.io/projected/0a40ce6c-6834-40bd-aad5-a34773f4e91b-kube-api-access-474fx\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100535 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-ca\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100551 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-encryption-config\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100566 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-service-ca-bundle\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100597 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-client-ca\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100637 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v762j\" (UniqueName: \"kubernetes.io/projected/d779a1a6-0749-48f9-a4a7-75e65a13c7da-kube-api-access-v762j\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100669 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sdqr\" (UniqueName: \"kubernetes.io/projected/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-kube-api-access-7sdqr\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100683 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-console-config\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100697 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jd8r\" (UniqueName: \"kubernetes.io/projected/18c06562-74d0-4b4c-be8e-2fcbe69c562b-kube-api-access-9jd8r\") pod \"migrator-59844c95c7-4kcl7\" (UID: \"18c06562-74d0-4b4c-be8e-2fcbe69c562b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100712 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-default-certificate\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100728 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2dqr\" (UniqueName: \"kubernetes.io/projected/15b7bb22-0ae1-46af-a1cf-4024996c996f-kube-api-access-q2dqr\") pod \"control-plane-machine-set-operator-78cbb6b69f-8wmkg\" (UID: \"15b7bb22-0ae1-46af-a1cf-4024996c996f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100760 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dbd3b91b-be30-474a-94d4-b4a684eef907-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100776 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10534d55-c886-4f76-b7ba-011bee2b965d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100793 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100807 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/33f93332-da64-4f57-b6cd-c4e643976bed-machine-approver-tls\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100838 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f93332-da64-4f57-b6cd-c4e643976bed-config\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100854 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100871 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-trusted-ca-bundle\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100904 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-policies\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100919 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a40ce6c-6834-40bd-aad5-a34773f4e91b-metrics-tls\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100937 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100951 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-serving-cert\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100981 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxbs8\" (UniqueName: \"kubernetes.io/projected/2c18bce1-7e19-4745-a51d-de4a04e44dd0-kube-api-access-dxbs8\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.100998 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5c2p\" (UniqueName: \"kubernetes.io/projected/de33dc58-134b-4791-9b40-7529ae2c8c8b-kube-api-access-l5c2p\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101014 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbd3b91b-be30-474a-94d4-b4a684eef907-serving-cert\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101029 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz67n\" (UniqueName: \"kubernetes.io/projected/b9b0d1de-c306-49d7-a94e-c83c054e53ef-kube-api-access-cz67n\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101064 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101081 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bfdc21f9-9837-4ab1-b89f-51d630bfc302-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101098 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101115 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6s8q\" (UniqueName: \"kubernetes.io/projected/f8096c4b-d1ba-471a-9d05-22aab449b51c-kube-api-access-g6s8q\") pod \"cluster-samples-operator-665b6dd947-fgr9n\" (UID: \"f8096c4b-d1ba-471a-9d05-22aab449b51c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101148 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8f74109-ec7a-423c-9330-8089c7935d18-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101164 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b09198a-0a36-412a-86f9-f723fb6a483c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101179 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101194 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101223 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-config\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101239 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-etcd-client\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101253 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101287 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101301 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b09198a-0a36-412a-86f9-f723fb6a483c-config\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101315 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfdc21f9-9837-4ab1-b89f-51d630bfc302-config\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101330 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-config\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101360 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101378 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dbk\" (UniqueName: \"kubernetes.io/projected/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-kube-api-access-d8dbk\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101394 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a40ce6c-6834-40bd-aad5-a34773f4e91b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101407 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-config\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101438 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/de33dc58-134b-4791-9b40-7529ae2c8c8b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101453 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/15b7bb22-0ae1-46af-a1cf-4024996c996f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8wmkg\" (UID: \"15b7bb22-0ae1-46af-a1cf-4024996c996f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101467 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-srv-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101484 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101513 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9d82114-9bc6-4bce-8418-46c860542aa4-audit-dir\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101530 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d779a1a6-0749-48f9-a4a7-75e65a13c7da-serving-cert\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101546 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxb52\" (UniqueName: \"kubernetes.io/projected/10534d55-c886-4f76-b7ba-011bee2b965d-kube-api-access-vxb52\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101560 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101574 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101604 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9p5r\" (UniqueName: \"kubernetes.io/projected/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-kube-api-access-z9p5r\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101619 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101635 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-oauth-serving-cert\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101652 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101682 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9b0d1de-c306-49d7-a94e-c83c054e53ef-service-ca-bundle\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101698 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101713 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f93332-da64-4f57-b6cd-c4e643976bed-auth-proxy-config\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101726 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-metrics-certs\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101759 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101780 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f74109-ec7a-423c-9330-8089c7935d18-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101796 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-images\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101831 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101855 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10534d55-c886-4f76-b7ba-011bee2b965d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101869 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f74109-ec7a-423c-9330-8089c7935d18-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101882 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a40ce6c-6834-40bd-aad5-a34773f4e91b-trusted-ca\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101916 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp8rd\" (UniqueName: \"kubernetes.io/projected/ce3173af-2480-4a96-a7c5-ab9c44b18c64-kube-api-access-hp8rd\") pod \"downloads-7954f5f757-rdzj8\" (UID: \"ce3173af-2480-4a96-a7c5-ab9c44b18c64\") " pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101931 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql44q\" (UniqueName: \"kubernetes.io/projected/33f93332-da64-4f57-b6cd-c4e643976bed-kube-api-access-ql44q\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101946 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-images\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.101982 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-config\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102003 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-serving-cert\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102029 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d4pz\" (UniqueName: \"kubernetes.io/projected/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-kube-api-access-4d4pz\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102071 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-audit-policies\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102091 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102111 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccrj7\" (UniqueName: \"kubernetes.io/projected/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-kube-api-access-ccrj7\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102144 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-dir\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102160 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102180 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts2d7\" (UniqueName: \"kubernetes.io/projected/dbd3b91b-be30-474a-94d4-b4a684eef907-kube-api-access-ts2d7\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102328 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2mbh\" (UniqueName: \"kubernetes.io/projected/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-kube-api-access-l2mbh\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102346 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-service-ca\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102381 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-service-ca\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102398 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b09198a-0a36-412a-86f9-f723fb6a483c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102413 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c18bce1-7e19-4745-a51d-de4a04e44dd0-serving-cert\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102427 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.102458 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.103107 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-config\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.103401 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-config\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.103496 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.103524 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-images\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.103604 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.103650 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9d82114-9bc6-4bce-8418-46c860542aa4-audit-dir\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104023 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-audit-policies\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104453 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9d82114-9bc6-4bce-8418-46c860542aa4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104605 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8096c4b-d1ba-471a-9d05-22aab449b51c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fgr9n\" (UID: \"f8096c4b-d1ba-471a-9d05-22aab449b51c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104632 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv6l4\" (UniqueName: \"kubernetes.io/projected/7aa958d0-0f0b-46c2-884f-e7f23431c022-kube-api-access-zv6l4\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104650 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/de33dc58-134b-4791-9b40-7529ae2c8c8b-proxy-tls\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104682 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mnk4\" (UniqueName: \"kubernetes.io/projected/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-kube-api-access-9mnk4\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104701 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-proxy-tls\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104717 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8z87\" (UniqueName: \"kubernetes.io/projected/dc437d52-8b8b-430f-bf9e-67492c365aad-kube-api-access-k8z87\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104737 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-config\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104745 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-stats-auth\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104764 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dbd3b91b-be30-474a-94d4-b4a684eef907-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104801 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-client\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.104831 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfdc21f9-9837-4ab1-b89f-51d630bfc302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.106656 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.107371 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-client-ca\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.107721 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d779a1a6-0749-48f9-a4a7-75e65a13c7da-service-ca-bundle\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.108417 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-etcd-client\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.108425 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-encryption-config\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.109241 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.109460 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.109690 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.109715 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.110880 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.112170 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d779a1a6-0749-48f9-a4a7-75e65a13c7da-serving-cert\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.112557 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9d82114-9bc6-4bce-8418-46c860542aa4-serving-cert\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.121180 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.122153 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8096c4b-d1ba-471a-9d05-22aab449b51c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fgr9n\" (UID: \"f8096c4b-d1ba-471a-9d05-22aab449b51c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.125920 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbd3b91b-be30-474a-94d4-b4a684eef907-serving-cert\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.127365 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.127897 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.127937 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-serving-cert\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128144 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128240 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128370 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128474 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128375 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128650 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128790 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.128416 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.131741 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pmhfw"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.131986 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.135770 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.147075 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.149028 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.149198 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.149511 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z8hd2"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.160389 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sngfr"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.160446 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.166084 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.166616 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.167883 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.186214 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.188471 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fdwzj"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.190439 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.191432 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.192952 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ktgzm"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.196690 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.198676 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.200007 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.202593 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.203003 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.204551 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zbnzg"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206650 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10534d55-c886-4f76-b7ba-011bee2b965d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206677 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206697 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/33f93332-da64-4f57-b6cd-c4e643976bed-machine-approver-tls\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206715 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f93332-da64-4f57-b6cd-c4e643976bed-config\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206733 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-trusted-ca-bundle\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206749 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-policies\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206767 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-serving-cert\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a40ce6c-6834-40bd-aad5-a34773f4e91b-metrics-tls\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206834 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxbs8\" (UniqueName: \"kubernetes.io/projected/2c18bce1-7e19-4745-a51d-de4a04e44dd0-kube-api-access-dxbs8\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206853 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5c2p\" (UniqueName: \"kubernetes.io/projected/de33dc58-134b-4791-9b40-7529ae2c8c8b-kube-api-access-l5c2p\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206871 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz67n\" (UniqueName: \"kubernetes.io/projected/b9b0d1de-c306-49d7-a94e-c83c054e53ef-kube-api-access-cz67n\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206886 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bfdc21f9-9837-4ab1-b89f-51d630bfc302-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206919 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.206937 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b09198a-0a36-412a-86f9-f723fb6a483c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208224 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208628 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208663 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8f74109-ec7a-423c-9330-8089c7935d18-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208682 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b09198a-0a36-412a-86f9-f723fb6a483c-config\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208716 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208734 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfdc21f9-9837-4ab1-b89f-51d630bfc302-config\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208752 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208774 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dbk\" (UniqueName: \"kubernetes.io/projected/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-kube-api-access-d8dbk\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208795 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/de33dc58-134b-4791-9b40-7529ae2c8c8b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208811 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/15b7bb22-0ae1-46af-a1cf-4024996c996f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8wmkg\" (UID: \"15b7bb22-0ae1-46af-a1cf-4024996c996f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208831 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a40ce6c-6834-40bd-aad5-a34773f4e91b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208856 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-config\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208878 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxb52\" (UniqueName: \"kubernetes.io/projected/10534d55-c886-4f76-b7ba-011bee2b965d-kube-api-access-vxb52\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208895 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-srv-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208913 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9p5r\" (UniqueName: \"kubernetes.io/projected/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-kube-api-access-z9p5r\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208931 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208949 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208966 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208984 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-oauth-serving-cert\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209002 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209021 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9b0d1de-c306-49d7-a94e-c83c054e53ef-service-ca-bundle\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209038 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f93332-da64-4f57-b6cd-c4e643976bed-auth-proxy-config\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209071 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-metrics-certs\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209105 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f74109-ec7a-423c-9330-8089c7935d18-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-images\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209136 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209150 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-policies\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209154 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10534d55-c886-4f76-b7ba-011bee2b965d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209205 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a40ce6c-6834-40bd-aad5-a34773f4e91b-trusted-ca\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209225 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f74109-ec7a-423c-9330-8089c7935d18-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209251 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql44q\" (UniqueName: \"kubernetes.io/projected/33f93332-da64-4f57-b6cd-c4e643976bed-kube-api-access-ql44q\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209306 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccrj7\" (UniqueName: \"kubernetes.io/projected/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-kube-api-access-ccrj7\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209329 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-dir\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209350 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209381 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-service-ca\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209397 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-service-ca\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209411 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b09198a-0a36-412a-86f9-f723fb6a483c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209425 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c18bce1-7e19-4745-a51d-de4a04e44dd0-serving-cert\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209439 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209454 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209471 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv6l4\" (UniqueName: \"kubernetes.io/projected/7aa958d0-0f0b-46c2-884f-e7f23431c022-kube-api-access-zv6l4\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209497 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-proxy-tls\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209511 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/de33dc58-134b-4791-9b40-7529ae2c8c8b-proxy-tls\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209533 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-stats-auth\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209546 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-client\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfdc21f9-9837-4ab1-b89f-51d630bfc302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209576 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8z87\" (UniqueName: \"kubernetes.io/projected/dc437d52-8b8b-430f-bf9e-67492c365aad-kube-api-access-k8z87\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209584 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-trusted-ca-bundle\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209598 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr7bs\" (UniqueName: \"kubernetes.io/projected/222d84fa-7732-4a9d-b250-e66c95092ad6-kube-api-access-lr7bs\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209654 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209685 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-oauth-config\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209713 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-474fx\" (UniqueName: \"kubernetes.io/projected/0a40ce6c-6834-40bd-aad5-a34773f4e91b-kube-api-access-474fx\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209733 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-ca\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209753 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209794 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-console-config\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209802 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-dir\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209813 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jd8r\" (UniqueName: \"kubernetes.io/projected/18c06562-74d0-4b4c-be8e-2fcbe69c562b-kube-api-access-9jd8r\") pod \"migrator-59844c95c7-4kcl7\" (UID: \"18c06562-74d0-4b4c-be8e-2fcbe69c562b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209836 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sdqr\" (UniqueName: \"kubernetes.io/projected/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-kube-api-access-7sdqr\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209855 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-default-certificate\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.209874 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2dqr\" (UniqueName: \"kubernetes.io/projected/15b7bb22-0ae1-46af-a1cf-4024996c996f-kube-api-access-q2dqr\") pod \"control-plane-machine-set-operator-78cbb6b69f-8wmkg\" (UID: \"15b7bb22-0ae1-46af-a1cf-4024996c996f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.210141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.210213 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.210466 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5brmj"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.210509 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.210519 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-2vdl8"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.210783 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-service-ca\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.211242 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/33f93332-da64-4f57-b6cd-c4e643976bed-machine-approver-tls\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.212694 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.213197 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f93332-da64-4f57-b6cd-c4e643976bed-auth-proxy-config\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.213662 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.214309 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-serving-cert\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.208755 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f93332-da64-4f57-b6cd-c4e643976bed-config\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.215104 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.215548 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.215723 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.215825 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.216142 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.216494 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/de33dc58-134b-4791-9b40-7529ae2c8c8b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.217186 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-console-config\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.217667 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-oauth-config\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.218531 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.219333 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-9dtg7"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.219629 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.219664 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.220345 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.220629 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.221020 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-oauth-serving-cert\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.221385 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.221730 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.222526 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.222779 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.223844 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.225541 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.228142 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gqch4"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.229430 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.229627 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.231934 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7rks7"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.232751 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7t7jg"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.233529 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/15b7bb22-0ae1-46af-a1cf-4024996c996f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8wmkg\" (UID: \"15b7bb22-0ae1-46af-a1cf-4024996c996f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.233635 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.234634 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2vdl8"] Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.244014 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.259378 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.279580 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.284170 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c18bce1-7e19-4745-a51d-de4a04e44dd0-serving-cert\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.299570 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.308142 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-client\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.319286 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.326061 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-config\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.339909 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.347443 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-ca\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.361435 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.371480 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c18bce1-7e19-4745-a51d-de4a04e44dd0-etcd-service-ca\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.380051 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.400218 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.420638 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.434848 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f74109-ec7a-423c-9330-8089c7935d18-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.440568 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.445036 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f74109-ec7a-423c-9330-8089c7935d18-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.463671 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.489538 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.499885 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.519379 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.540015 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.560070 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.570986 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10534d55-c886-4f76-b7ba-011bee2b965d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.580630 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.600958 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.610057 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10534d55-c886-4f76-b7ba-011bee2b965d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.620333 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.642125 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.661029 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.680927 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.701002 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.714116 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a40ce6c-6834-40bd-aad5-a34773f4e91b-metrics-tls\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.722980 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.755833 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.759738 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.762261 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a40ce6c-6834-40bd-aad5-a34773f4e91b-trusted-ca\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.779987 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.788797 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/de33dc58-134b-4791-9b40-7529ae2c8c8b-proxy-tls\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.805456 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.819308 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.829979 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-images\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.840519 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.860235 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.880002 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.888331 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-proxy-tls\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.901007 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.914350 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b09198a-0a36-412a-86f9-f723fb6a483c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.919723 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.926064 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b09198a-0a36-412a-86f9-f723fb6a483c-config\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.940406 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.960764 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.968488 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-metrics-certs\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.980391 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 11:27:23 crc kubenswrapper[4796]: I1127 11:27:23.989596 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-stats-auth\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.000419 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.006752 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b9b0d1de-c306-49d7-a94e-c83c054e53ef-default-certificate\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.020675 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.038734 4796 request.go:700] Waited for 1.019380754s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dservice-ca-bundle&limit=500&resourceVersion=0 Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.040075 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.043566 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9b0d1de-c306-49d7-a94e-c83c054e53ef-service-ca-bundle\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.060585 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.080741 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.100224 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.128742 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.137584 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.140353 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.161708 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.175393 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.181689 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.200256 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.212972 4796 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.213056 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-profile-collector-cert podName:3e6acd38-f2b9-4628-8d87-c269ef09d9bc nodeName:}" failed. No retries permitted until 2025-11-27 11:27:24.713029689 +0000 UTC m=+162.231348637 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-profile-collector-cert") pod "catalog-operator-68c6474976-cwptf" (UID: "3e6acd38-f2b9-4628-8d87-c269ef09d9bc") : failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.213166 4796 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.213223 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-srv-cert podName:3e6acd38-f2b9-4628-8d87-c269ef09d9bc nodeName:}" failed. No retries permitted until 2025-11-27 11:27:24.713205703 +0000 UTC m=+162.231524661 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-srv-cert") pod "catalog-operator-68c6474976-cwptf" (UID: "3e6acd38-f2b9-4628-8d87-c269ef09d9bc") : failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.214557 4796 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.214623 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-webhook-certs podName:0d8269c0-b923-4ad0-8aa8-ddb2e21b257d nodeName:}" failed. No retries permitted until 2025-11-27 11:27:24.714608509 +0000 UTC m=+162.232927457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-webhook-certs") pod "multus-admission-controller-857f4d67dd-7rks7" (UID: "0d8269c0-b923-4ad0-8aa8-ddb2e21b257d") : failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.215798 4796 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.215904 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfdc21f9-9837-4ab1-b89f-51d630bfc302-serving-cert podName:bfdc21f9-9837-4ab1-b89f-51d630bfc302 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:24.715873852 +0000 UTC m=+162.234192810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/bfdc21f9-9837-4ab1-b89f-51d630bfc302-serving-cert") pod "kube-controller-manager-operator-78b949d7b-4lbbd" (UID: "bfdc21f9-9837-4ab1-b89f-51d630bfc302") : failed to sync secret cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.216014 4796 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: E1127 11:27:24.216159 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bfdc21f9-9837-4ab1-b89f-51d630bfc302-config podName:bfdc21f9-9837-4ab1-b89f-51d630bfc302 nodeName:}" failed. No retries permitted until 2025-11-27 11:27:24.716125238 +0000 UTC m=+162.234444196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/bfdc21f9-9837-4ab1-b89f-51d630bfc302-config") pod "kube-controller-manager-operator-78b949d7b-4lbbd" (UID: "bfdc21f9-9837-4ab1-b89f-51d630bfc302") : failed to sync configmap cache: timed out waiting for the condition Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.220651 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.242762 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.260736 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.300003 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.321167 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.340420 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.360611 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.380094 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.400598 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.420405 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.460933 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.480257 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.500085 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.521859 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.541869 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.560662 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.580668 4796 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.608546 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.621259 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.641958 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.660731 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.681992 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.699473 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.722930 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.736048 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.736096 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfdc21f9-9837-4ab1-b89f-51d630bfc302-config\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.736162 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-srv-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.736202 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.736381 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfdc21f9-9837-4ab1-b89f-51d630bfc302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.737505 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfdc21f9-9837-4ab1-b89f-51d630bfc302-config\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.741007 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.741207 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-srv-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.742007 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.750550 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfdc21f9-9837-4ab1-b89f-51d630bfc302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.755203 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.760636 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.780603 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.800721 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.820006 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.840423 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.860509 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.895751 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.941140 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d4pz\" (UniqueName: \"kubernetes.io/projected/96d741c5-ebc0-4c04-9192-c2661a5d0e0a-kube-api-access-4d4pz\") pod \"machine-api-operator-5694c8668f-bxf5l\" (UID: \"96d741c5-ebc0-4c04-9192-c2661a5d0e0a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.950005 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zq8t\" (UniqueName: \"kubernetes.io/projected/f9d82114-9bc6-4bce-8418-46c860542aa4-kube-api-access-7zq8t\") pod \"apiserver-7bbb656c7d-hlhhp\" (UID: \"f9d82114-9bc6-4bce-8418-46c860542aa4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.970666 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp8rd\" (UniqueName: \"kubernetes.io/projected/ce3173af-2480-4a96-a7c5-ab9c44b18c64-kube-api-access-hp8rd\") pod \"downloads-7954f5f757-rdzj8\" (UID: \"ce3173af-2480-4a96-a7c5-ab9c44b18c64\") " pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.975735 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts2d7\" (UniqueName: \"kubernetes.io/projected/dbd3b91b-be30-474a-94d4-b4a684eef907-kube-api-access-ts2d7\") pod \"openshift-config-operator-7777fb866f-zs7tf\" (UID: \"dbd3b91b-be30-474a-94d4-b4a684eef907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:24 crc kubenswrapper[4796]: I1127 11:27:24.996399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2mbh\" (UniqueName: \"kubernetes.io/projected/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-kube-api-access-l2mbh\") pod \"route-controller-manager-6576b87f9c-l8bvc\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.025801 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6s8q\" (UniqueName: \"kubernetes.io/projected/f8096c4b-d1ba-471a-9d05-22aab449b51c-kube-api-access-g6s8q\") pod \"cluster-samples-operator-665b6dd947-fgr9n\" (UID: \"f8096c4b-d1ba-471a-9d05-22aab449b51c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.038853 4796 request.go:700] Waited for 1.93101149s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.047524 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v762j\" (UniqueName: \"kubernetes.io/projected/d779a1a6-0749-48f9-a4a7-75e65a13c7da-kube-api-access-v762j\") pod \"authentication-operator-69f744f599-skdgb\" (UID: \"d779a1a6-0749-48f9-a4a7-75e65a13c7da\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.092852 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxbs8\" (UniqueName: \"kubernetes.io/projected/2c18bce1-7e19-4745-a51d-de4a04e44dd0-kube-api-access-dxbs8\") pod \"etcd-operator-b45778765-z8hd2\" (UID: \"2c18bce1-7e19-4745-a51d-de4a04e44dd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.098793 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.108413 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.123860 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5c2p\" (UniqueName: \"kubernetes.io/projected/de33dc58-134b-4791-9b40-7529ae2c8c8b-kube-api-access-l5c2p\") pod \"machine-config-controller-84d6567774-2bbpf\" (UID: \"de33dc58-134b-4791-9b40-7529ae2c8c8b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.132366 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mnk4\" (UniqueName: \"kubernetes.io/projected/e7f8ca38-34d4-4e26-b1e5-8798f9986d4c-kube-api-access-9mnk4\") pod \"openshift-apiserver-operator-796bbdcf4f-lsx6g\" (UID: \"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.138227 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.149923 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr7bs\" (UniqueName: \"kubernetes.io/projected/222d84fa-7732-4a9d-b250-e66c95092ad6-kube-api-access-lr7bs\") pod \"console-f9d7485db-bvccg\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.150337 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz67n\" (UniqueName: \"kubernetes.io/projected/b9b0d1de-c306-49d7-a94e-c83c054e53ef-kube-api-access-cz67n\") pod \"router-default-5444994796-v65j9\" (UID: \"b9b0d1de-c306-49d7-a94e-c83c054e53ef\") " pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.160309 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.162838 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.177139 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:25 crc kubenswrapper[4796]: W1127 11:27:25.181331 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9b0d1de_c306_49d7_a94e_c83c054e53ef.slice/crio-918b944fc7c7523a73504ba3b8af7fbc00084ab336d9f75032d997c30714cd9e WatchSource:0}: Error finding container 918b944fc7c7523a73504ba3b8af7fbc00084ab336d9f75032d997c30714cd9e: Status 404 returned error can't find the container with id 918b944fc7c7523a73504ba3b8af7fbc00084ab336d9f75032d997c30714cd9e Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.183212 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql44q\" (UniqueName: \"kubernetes.io/projected/33f93332-da64-4f57-b6cd-c4e643976bed-kube-api-access-ql44q\") pod \"machine-approver-56656f9798-8qplk\" (UID: \"33f93332-da64-4f57-b6cd-c4e643976bed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.196075 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccrj7\" (UniqueName: \"kubernetes.io/projected/8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e-kube-api-access-ccrj7\") pod \"machine-config-operator-74547568cd-x97jq\" (UID: \"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.207065 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2dqr\" (UniqueName: \"kubernetes.io/projected/15b7bb22-0ae1-46af-a1cf-4024996c996f-kube-api-access-q2dqr\") pod \"control-plane-machine-set-operator-78cbb6b69f-8wmkg\" (UID: \"15b7bb22-0ae1-46af-a1cf-4024996c996f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.209502 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.216536 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bfdc21f9-9837-4ab1-b89f-51d630bfc302-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4lbbd\" (UID: \"bfdc21f9-9837-4ab1-b89f-51d630bfc302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.239894 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.256689 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b09198a-0a36-412a-86f9-f723fb6a483c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kvksm\" (UID: \"5b09198a-0a36-412a-86f9-f723fb6a483c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.256841 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxb52\" (UniqueName: \"kubernetes.io/projected/10534d55-c886-4f76-b7ba-011bee2b965d-kube-api-access-vxb52\") pod \"kube-storage-version-migrator-operator-b67b599dd-8tdzt\" (UID: \"10534d55-c886-4f76-b7ba-011bee2b965d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.279178 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.290723 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.291982 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sdqr\" (UniqueName: \"kubernetes.io/projected/0d8269c0-b923-4ad0-8aa8-ddb2e21b257d-kube-api-access-7sdqr\") pod \"multus-admission-controller-857f4d67dd-7rks7\" (UID: \"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.300740 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jd8r\" (UniqueName: \"kubernetes.io/projected/18c06562-74d0-4b4c-be8e-2fcbe69c562b-kube-api-access-9jd8r\") pod \"migrator-59844c95c7-4kcl7\" (UID: \"18c06562-74d0-4b4c-be8e-2fcbe69c562b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.322071 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9p5r\" (UniqueName: \"kubernetes.io/projected/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-kube-api-access-z9p5r\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.336943 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4726245-f4f8-494f-b2d9-2d8cb83f05e3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-85t6r\" (UID: \"b4726245-f4f8-494f-b2d9-2d8cb83f05e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.341197 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.365342 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.386301 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.387091 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8f74109-ec7a-423c-9330-8089c7935d18-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mz8zx\" (UID: \"d8f74109-ec7a-423c-9330-8089c7935d18\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.392394 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.399030 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.400171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a40ce6c-6834-40bd-aad5-a34773f4e91b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.400467 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.413224 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" Nov 27 11:27:25 crc kubenswrapper[4796]: W1127 11:27:25.414333 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33f93332_da64_4f57_b6cd_c4e643976bed.slice/crio-2c28ba551a737a409aab05375b7a815827b8972a38d6cf7f0cbceb8b9676a6e5 WatchSource:0}: Error finding container 2c28ba551a737a409aab05375b7a815827b8972a38d6cf7f0cbceb8b9676a6e5: Status 404 returned error can't find the container with id 2c28ba551a737a409aab05375b7a815827b8972a38d6cf7f0cbceb8b9676a6e5 Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.420243 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.421338 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.423695 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.427456 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.444938 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.452171 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.456689 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.459906 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv6l4\" (UniqueName: \"kubernetes.io/projected/7aa958d0-0f0b-46c2-884f-e7f23431c022-kube-api-access-zv6l4\") pod \"oauth-openshift-558db77b4-sngfr\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.473192 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.479066 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dbk\" (UniqueName: \"kubernetes.io/projected/3e6acd38-f2b9-4628-8d87-c269ef09d9bc-kube-api-access-d8dbk\") pod \"catalog-operator-68c6474976-cwptf\" (UID: \"3e6acd38-f2b9-4628-8d87-c269ef09d9bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.480939 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.495617 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.501287 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-474fx\" (UniqueName: \"kubernetes.io/projected/0a40ce6c-6834-40bd-aad5-a34773f4e91b-kube-api-access-474fx\") pod \"ingress-operator-5b745b69d9-lxb9g\" (UID: \"0a40ce6c-6834-40bd-aad5-a34773f4e91b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.520115 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.533908 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8z87\" (UniqueName: \"kubernetes.io/projected/dc437d52-8b8b-430f-bf9e-67492c365aad-kube-api-access-k8z87\") pod \"marketplace-operator-79b997595-fdwzj\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.540397 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.561872 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.622682 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.646778 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwjvd\" (UniqueName: \"kubernetes.io/projected/aba9e15a-aeb7-40f8-8c19-e68305bf379e-kube-api-access-rwjvd\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.646825 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-profile-collector-cert\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.646851 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e1b281d-0c21-49d8-89fd-ec2534405cdf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.646890 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.646910 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.646935 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aba9e15a-aeb7-40f8-8c19-e68305bf379e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.647044 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-serving-cert\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: E1127 11:27:25.647749 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.14772885 +0000 UTC m=+163.666047808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.647797 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-image-import-ca\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.647832 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-tls\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.647894 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-872x8\" (UniqueName: \"kubernetes.io/projected/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-kube-api-access-872x8\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.647946 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j86dj\" (UniqueName: \"kubernetes.io/projected/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-kube-api-access-j86dj\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.648202 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-trusted-ca\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.648222 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmrx\" (UniqueName: \"kubernetes.io/projected/10b64bb3-de9d-4c66-8acc-13369ba5dc24-kube-api-access-kgmrx\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.648282 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-encryption-config\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.648308 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-serving-cert\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.648357 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-config\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.648545 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95abf6de-142f-47bc-9492-38fc1e8f69b0-audit-dir\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.650918 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gddqg\" (UniqueName: \"kubernetes.io/projected/43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb-kube-api-access-gddqg\") pod \"dns-operator-744455d44c-l954k\" (UID: \"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb\") " pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.650991 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e1b281d-0c21-49d8-89fd-ec2534405cdf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651025 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aba9e15a-aeb7-40f8-8c19-e68305bf379e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651064 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/95abf6de-142f-47bc-9492-38fc1e8f69b0-node-pullsecrets\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651108 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-srv-cert\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651132 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-etcd-serving-ca\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-client-ca\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651188 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-config\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651234 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-bound-sa-token\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651283 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b64bb3-de9d-4c66-8acc-13369ba5dc24-serving-cert\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651339 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb-metrics-tls\") pod \"dns-operator-744455d44c-l954k\" (UID: \"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb\") " pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651361 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b64bb3-de9d-4c66-8acc-13369ba5dc24-config\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651391 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-certificates\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651423 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-audit\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651447 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntqtr\" (UniqueName: \"kubernetes.io/projected/95abf6de-142f-47bc-9492-38fc1e8f69b0-kube-api-access-ntqtr\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651493 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-etcd-client\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651746 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651831 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/10b64bb3-de9d-4c66-8acc-13369ba5dc24-trusted-ca\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.651858 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzjww\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-kube-api-access-xzjww\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.659324 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.728857 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753100 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753218 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-socket-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753240 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thjhj\" (UniqueName: \"kubernetes.io/projected/0d08ea76-e06c-433f-986a-def48c08f46f-kube-api-access-thjhj\") pod \"package-server-manager-789f6589d5-pzxv6\" (UID: \"0d08ea76-e06c-433f-986a-def48c08f46f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753289 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-metrics-tls\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753361 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb-metrics-tls\") pod \"dns-operator-744455d44c-l954k\" (UID: \"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb\") " pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753377 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b64bb3-de9d-4c66-8acc-13369ba5dc24-config\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753410 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-certificates\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753434 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7abc2a6e-b355-4186-9dce-56f8ec89ef14-signing-key\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753466 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-audit\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753481 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntqtr\" (UniqueName: \"kubernetes.io/projected/95abf6de-142f-47bc-9492-38fc1e8f69b0-kube-api-access-ntqtr\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753515 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-etcd-client\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753550 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d08ea76-e06c-433f-986a-def48c08f46f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pzxv6\" (UID: \"0d08ea76-e06c-433f-986a-def48c08f46f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753573 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-tmpfs\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753609 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753626 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/10b64bb3-de9d-4c66-8acc-13369ba5dc24-trusted-ca\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753642 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t82bp\" (UniqueName: \"kubernetes.io/projected/39c826c0-9aa1-4fed-a8ca-3dfafe5f5365-kube-api-access-t82bp\") pod \"ingress-canary-gqch4\" (UID: \"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365\") " pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753658 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7aed61d7-778d-4820-881a-e087775da1a1-certs\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753700 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzjww\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-kube-api-access-xzjww\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753716 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7abc2a6e-b355-4186-9dce-56f8ec89ef14-signing-cabundle\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753732 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klcgc\" (UniqueName: \"kubernetes.io/projected/7aed61d7-778d-4820-881a-e087775da1a1-kube-api-access-klcgc\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753777 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e1b281d-0c21-49d8-89fd-ec2534405cdf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753850 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwjvd\" (UniqueName: \"kubernetes.io/projected/aba9e15a-aeb7-40f8-8c19-e68305bf379e-kube-api-access-rwjvd\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753867 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-profile-collector-cert\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753884 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k4v9\" (UniqueName: \"kubernetes.io/projected/517b90c8-7ceb-43fa-9d57-a2c5ac735138-kube-api-access-6k4v9\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753900 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-registration-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753914 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5mg2\" (UniqueName: \"kubernetes.io/projected/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-kube-api-access-j5mg2\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753944 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-config-volume\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.753959 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/517b90c8-7ceb-43fa-9d57-a2c5ac735138-serving-cert\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754004 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754030 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aba9e15a-aeb7-40f8-8c19-e68305bf379e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754075 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-serving-cert\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754108 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-image-import-ca\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754140 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-tls\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754157 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-872x8\" (UniqueName: \"kubernetes.io/projected/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-kube-api-access-872x8\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754173 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgv2\" (UniqueName: \"kubernetes.io/projected/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-kube-api-access-flgv2\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754217 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j86dj\" (UniqueName: \"kubernetes.io/projected/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-kube-api-access-j86dj\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754236 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-trusted-ca\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754251 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmrx\" (UniqueName: \"kubernetes.io/projected/10b64bb3-de9d-4c66-8acc-13369ba5dc24-kube-api-access-kgmrx\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754299 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-csi-data-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754334 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9297600d-7a17-4a87-9ed1-26357ea3df51-secret-volume\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754350 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7aed61d7-778d-4820-881a-e087775da1a1-node-bootstrap-token\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754384 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-encryption-config\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754409 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-webhook-cert\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754426 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9297600d-7a17-4a87-9ed1-26357ea3df51-config-volume\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754454 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-serving-cert\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754564 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-mountpoint-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754580 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp69x\" (UniqueName: \"kubernetes.io/projected/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-kube-api-access-dp69x\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754598 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-plugins-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-config\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754638 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95abf6de-142f-47bc-9492-38fc1e8f69b0-audit-dir\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754664 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gddqg\" (UniqueName: \"kubernetes.io/projected/43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb-kube-api-access-gddqg\") pod \"dns-operator-744455d44c-l954k\" (UID: \"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb\") " pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754680 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmlqg\" (UniqueName: \"kubernetes.io/projected/7abc2a6e-b355-4186-9dce-56f8ec89ef14-kube-api-access-pmlqg\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754736 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e1b281d-0c21-49d8-89fd-ec2534405cdf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754752 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39c826c0-9aa1-4fed-a8ca-3dfafe5f5365-cert\") pod \"ingress-canary-gqch4\" (UID: \"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365\") " pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754769 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aba9e15a-aeb7-40f8-8c19-e68305bf379e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754794 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/95abf6de-142f-47bc-9492-38fc1e8f69b0-node-pullsecrets\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754840 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z64vn\" (UniqueName: \"kubernetes.io/projected/9297600d-7a17-4a87-9ed1-26357ea3df51-kube-api-access-z64vn\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754857 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-apiservice-cert\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754947 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-srv-cert\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.754998 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-etcd-serving-ca\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.755015 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-client-ca\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.755031 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-config\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.755216 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-bound-sa-token\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.755309 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/517b90c8-7ceb-43fa-9d57-a2c5ac735138-config\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.755372 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b64bb3-de9d-4c66-8acc-13369ba5dc24-serving-cert\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.756241 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-image-import-ca\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.756619 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-etcd-serving-ca\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.757074 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-client-ca\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.757562 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-config\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.757719 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b64bb3-de9d-4c66-8acc-13369ba5dc24-config\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.758119 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-config\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.758182 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/95abf6de-142f-47bc-9492-38fc1e8f69b0-audit-dir\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.759297 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aba9e15a-aeb7-40f8-8c19-e68305bf379e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.759351 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/95abf6de-142f-47bc-9492-38fc1e8f69b0-node-pullsecrets\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: E1127 11:27:25.760157 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.260136397 +0000 UTC m=+163.778455365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.760795 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b64bb3-de9d-4c66-8acc-13369ba5dc24-serving-cert\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.760916 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-tls\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.760169 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e1b281d-0c21-49d8-89fd-ec2534405cdf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.762678 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/10b64bb3-de9d-4c66-8acc-13369ba5dc24-trusted-ca\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.762992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-etcd-client\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.763059 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-serving-cert\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.763466 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.764063 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-srv-cert\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.768402 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.768722 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aba9e15a-aeb7-40f8-8c19-e68305bf379e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.768997 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-serving-cert\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.769106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb-metrics-tls\") pod \"dns-operator-744455d44c-l954k\" (UID: \"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb\") " pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.769954 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e1b281d-0c21-49d8-89fd-ec2534405cdf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.770450 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/95abf6de-142f-47bc-9492-38fc1e8f69b0-encryption-config\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.770585 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-profile-collector-cert\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.805091 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.806094 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/95abf6de-142f-47bc-9492-38fc1e8f69b0-audit\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.806436 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-certificates\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.806650 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-trusted-ca\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.809666 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-872x8\" (UniqueName: \"kubernetes.io/projected/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-kube-api-access-872x8\") pod \"controller-manager-879f6c89f-pmhfw\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.818403 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gddqg\" (UniqueName: \"kubernetes.io/projected/43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb-kube-api-access-gddqg\") pod \"dns-operator-744455d44c-l954k\" (UID: \"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb\") " pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.819198 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.820724 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.822443 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bxf5l"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.852838 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntqtr\" (UniqueName: \"kubernetes.io/projected/95abf6de-142f-47bc-9492-38fc1e8f69b0-kube-api-access-ntqtr\") pod \"apiserver-76f77b778f-dp4q8\" (UID: \"95abf6de-142f-47bc-9492-38fc1e8f69b0\") " pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/517b90c8-7ceb-43fa-9d57-a2c5ac735138-config\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856644 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-socket-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856661 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thjhj\" (UniqueName: \"kubernetes.io/projected/0d08ea76-e06c-433f-986a-def48c08f46f-kube-api-access-thjhj\") pod \"package-server-manager-789f6589d5-pzxv6\" (UID: \"0d08ea76-e06c-433f-986a-def48c08f46f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856680 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-metrics-tls\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856701 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7abc2a6e-b355-4186-9dce-56f8ec89ef14-signing-key\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856720 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d08ea76-e06c-433f-986a-def48c08f46f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pzxv6\" (UID: \"0d08ea76-e06c-433f-986a-def48c08f46f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-tmpfs\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856753 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t82bp\" (UniqueName: \"kubernetes.io/projected/39c826c0-9aa1-4fed-a8ca-3dfafe5f5365-kube-api-access-t82bp\") pod \"ingress-canary-gqch4\" (UID: \"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365\") " pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856780 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7aed61d7-778d-4820-881a-e087775da1a1-certs\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856803 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7abc2a6e-b355-4186-9dce-56f8ec89ef14-signing-cabundle\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856817 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klcgc\" (UniqueName: \"kubernetes.io/projected/7aed61d7-778d-4820-881a-e087775da1a1-kube-api-access-klcgc\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856838 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k4v9\" (UniqueName: \"kubernetes.io/projected/517b90c8-7ceb-43fa-9d57-a2c5ac735138-kube-api-access-6k4v9\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856853 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-registration-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856866 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5mg2\" (UniqueName: \"kubernetes.io/projected/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-kube-api-access-j5mg2\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856883 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856898 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-config-volume\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856912 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/517b90c8-7ceb-43fa-9d57-a2c5ac735138-serving-cert\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856933 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flgv2\" (UniqueName: \"kubernetes.io/projected/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-kube-api-access-flgv2\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856961 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-csi-data-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856983 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9297600d-7a17-4a87-9ed1-26357ea3df51-secret-volume\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.856998 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7aed61d7-778d-4820-881a-e087775da1a1-node-bootstrap-token\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857014 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-webhook-cert\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857031 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9297600d-7a17-4a87-9ed1-26357ea3df51-config-volume\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857063 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-mountpoint-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857078 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp69x\" (UniqueName: \"kubernetes.io/projected/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-kube-api-access-dp69x\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857094 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-plugins-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857170 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmlqg\" (UniqueName: \"kubernetes.io/projected/7abc2a6e-b355-4186-9dce-56f8ec89ef14-kube-api-access-pmlqg\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857184 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39c826c0-9aa1-4fed-a8ca-3dfafe5f5365-cert\") pod \"ingress-canary-gqch4\" (UID: \"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365\") " pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857203 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z64vn\" (UniqueName: \"kubernetes.io/projected/9297600d-7a17-4a87-9ed1-26357ea3df51-kube-api-access-z64vn\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.857220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-apiservice-cert\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: E1127 11:27:25.857907 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.357892099 +0000 UTC m=+163.876211017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.858115 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-tmpfs\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.858907 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/517b90c8-7ceb-43fa-9d57-a2c5ac735138-config\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.861664 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-mountpoint-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.861790 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-plugins-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.861915 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-socket-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.862091 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9297600d-7a17-4a87-9ed1-26357ea3df51-config-volume\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.862822 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-registration-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.864505 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-metrics-tls\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.864505 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-csi-data-dir\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.864848 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/517b90c8-7ceb-43fa-9d57-a2c5ac735138-serving-cert\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.865238 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7abc2a6e-b355-4186-9dce-56f8ec89ef14-signing-key\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.865640 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-apiservice-cert\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.866193 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j86dj\" (UniqueName: \"kubernetes.io/projected/fc25bf6e-f7fc-4e32-a802-ba3df37f6688-kube-api-access-j86dj\") pod \"olm-operator-6b444d44fb-g949p\" (UID: \"fc25bf6e-f7fc-4e32-a802-ba3df37f6688\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.866198 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" event={"ID":"dbd3b91b-be30-474a-94d4-b4a684eef907","Type":"ContainerStarted","Data":"5175b8ddfc88335de1b40c379381a532c9b0608ca7c643a0256b4be5477ea8ef"} Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.867576 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" event={"ID":"f9d82114-9bc6-4bce-8418-46c860542aa4","Type":"ContainerStarted","Data":"429dbf9cab9287acf7e3be87eb758c8162f0d2a3eb20805b008922fb4e47a599"} Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.868216 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/39c826c0-9aa1-4fed-a8ca-3dfafe5f5365-cert\") pod \"ingress-canary-gqch4\" (UID: \"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365\") " pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.871525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9297600d-7a17-4a87-9ed1-26357ea3df51-secret-volume\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.873130 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d08ea76-e06c-433f-986a-def48c08f46f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pzxv6\" (UID: \"0d08ea76-e06c-433f-986a-def48c08f46f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.874491 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7aed61d7-778d-4820-881a-e087775da1a1-node-bootstrap-token\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.877821 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-webhook-cert\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.878482 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-v65j9" event={"ID":"b9b0d1de-c306-49d7-a94e-c83c054e53ef","Type":"ContainerStarted","Data":"2382f2f34ee59efd8d497e6a636a8082634b4ba9c7d91d45affef7c1c421ab7f"} Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.878516 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-v65j9" event={"ID":"b9b0d1de-c306-49d7-a94e-c83c054e53ef","Type":"ContainerStarted","Data":"918b944fc7c7523a73504ba3b8af7fbc00084ab336d9f75032d997c30714cd9e"} Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.880986 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" event={"ID":"33f93332-da64-4f57-b6cd-c4e643976bed","Type":"ContainerStarted","Data":"2c28ba551a737a409aab05375b7a815827b8972a38d6cf7f0cbceb8b9676a6e5"} Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.884291 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmrx\" (UniqueName: \"kubernetes.io/projected/10b64bb3-de9d-4c66-8acc-13369ba5dc24-kube-api-access-kgmrx\") pod \"console-operator-58897d9998-zbnzg\" (UID: \"10b64bb3-de9d-4c66-8acc-13369ba5dc24\") " pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.897355 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.898118 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-bound-sa-token\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.905639 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-config-volume\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.905945 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7aed61d7-778d-4820-881a-e087775da1a1-certs\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.908373 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7abc2a6e-b355-4186-9dce-56f8ec89ef14-signing-cabundle\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.916122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzjww\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-kube-api-access-xzjww\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.950018 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwjvd\" (UniqueName: \"kubernetes.io/projected/aba9e15a-aeb7-40f8-8c19-e68305bf379e-kube-api-access-rwjvd\") pod \"openshift-controller-manager-operator-756b6f6bc6-4fq6z\" (UID: \"aba9e15a-aeb7-40f8-8c19-e68305bf379e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.951862 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.963203 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:25 crc kubenswrapper[4796]: E1127 11:27:25.963380 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.463357428 +0000 UTC m=+163.981676346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.963546 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:25 crc kubenswrapper[4796]: E1127 11:27:25.963926 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.463912942 +0000 UTC m=+163.982231860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.972540 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.979221 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l954k" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.985049 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rdzj8"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.986138 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-skdgb"] Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.989647 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t82bp\" (UniqueName: \"kubernetes.io/projected/39c826c0-9aa1-4fed-a8ca-3dfafe5f5365-kube-api-access-t82bp\") pod \"ingress-canary-gqch4\" (UID: \"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365\") " pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:25 crc kubenswrapper[4796]: I1127 11:27:25.994177 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgv2\" (UniqueName: \"kubernetes.io/projected/5f7f0a02-e94f-409c-8ed0-c5a7b301bace-kube-api-access-flgv2\") pod \"csi-hostpathplugin-7t7jg\" (UID: \"5f7f0a02-e94f-409c-8ed0-c5a7b301bace\") " pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.019452 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k4v9\" (UniqueName: \"kubernetes.io/projected/517b90c8-7ceb-43fa-9d57-a2c5ac735138-kube-api-access-6k4v9\") pod \"service-ca-operator-777779d784-gd7r5\" (UID: \"517b90c8-7ceb-43fa-9d57-a2c5ac735138\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.038158 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp69x\" (UniqueName: \"kubernetes.io/projected/cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6-kube-api-access-dp69x\") pod \"packageserver-d55dfcdfc-6nhkv\" (UID: \"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.054449 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmlqg\" (UniqueName: \"kubernetes.io/projected/7abc2a6e-b355-4186-9dce-56f8ec89ef14-kube-api-access-pmlqg\") pod \"service-ca-9c57cc56f-ktgzm\" (UID: \"7abc2a6e-b355-4186-9dce-56f8ec89ef14\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:26 crc kubenswrapper[4796]: W1127 11:27:26.060834 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce3173af_2480_4a96_a7c5_ab9c44b18c64.slice/crio-6d9867c124cf8e6284c300546fc2dc7df2e530f65591a01398cc6476c54f73d4 WatchSource:0}: Error finding container 6d9867c124cf8e6284c300546fc2dc7df2e530f65591a01398cc6476c54f73d4: Status 404 returned error can't find the container with id 6d9867c124cf8e6284c300546fc2dc7df2e530f65591a01398cc6476c54f73d4 Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.064826 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.064979 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.564957907 +0000 UTC m=+164.083276815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.065051 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.065406 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.565397818 +0000 UTC m=+164.083716736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: W1127 11:27:26.066207 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd779a1a6_0749_48f9_a4a7_75e65a13c7da.slice/crio-f454b537ff1edd0ac501de79989b8d547d29f732838d0625984b6be68e1ac5b1 WatchSource:0}: Error finding container f454b537ff1edd0ac501de79989b8d547d29f732838d0625984b6be68e1ac5b1: Status 404 returned error can't find the container with id f454b537ff1edd0ac501de79989b8d547d29f732838d0625984b6be68e1ac5b1 Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.074298 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klcgc\" (UniqueName: \"kubernetes.io/projected/7aed61d7-778d-4820-881a-e087775da1a1-kube-api-access-klcgc\") pod \"machine-config-server-9dtg7\" (UID: \"7aed61d7-778d-4820-881a-e087775da1a1\") " pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.086787 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.095771 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thjhj\" (UniqueName: \"kubernetes.io/projected/0d08ea76-e06c-433f-986a-def48c08f46f-kube-api-access-thjhj\") pod \"package-server-manager-789f6589d5-pzxv6\" (UID: \"0d08ea76-e06c-433f-986a-def48c08f46f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.115041 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5mg2\" (UniqueName: \"kubernetes.io/projected/ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3-kube-api-access-j5mg2\") pod \"dns-default-2vdl8\" (UID: \"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3\") " pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.128348 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.135799 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z64vn\" (UniqueName: \"kubernetes.io/projected/9297600d-7a17-4a87-9ed1-26357ea3df51-kube-api-access-z64vn\") pod \"collect-profiles-29404035-jj5kk\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.136058 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.143317 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.150287 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.158184 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.165860 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.165969 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.665951111 +0000 UTC m=+164.184270029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.166120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.166473 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.666463754 +0000 UTC m=+164.184782732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.168586 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.177524 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gqch4" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.192955 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.198437 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-9dtg7" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.246197 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.270357 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.270829 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.770808284 +0000 UTC m=+164.289127202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.371555 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.371988 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.871968374 +0000 UTC m=+164.390287292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.462480 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.470732 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.472816 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.473322 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:26.973289176 +0000 UTC m=+164.491608094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.475252 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.504914 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.504969 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bvccg"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.510155 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z8hd2"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.511903 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.513629 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.516662 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7rks7"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.520352 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.526293 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.526325 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.578606 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.580135 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.08011717 +0000 UTC m=+164.598436108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.679893 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.680198 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.18018028 +0000 UTC m=+164.698499208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.680642 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.680918 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.180910199 +0000 UTC m=+164.699229117 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.731421 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.737393 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sngfr"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.741776 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fdwzj"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.743898 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dp4q8"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.754304 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zbnzg"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.756053 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.757630 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.783982 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.784539 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.28451766 +0000 UTC m=+164.802836578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.886314 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.886830 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" event={"ID":"f8096c4b-d1ba-471a-9d05-22aab449b51c","Type":"ContainerStarted","Data":"295242d1217580da1b4fb585b7736082e3c69c439c9a24da23125fced1be489a"} Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.886946 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.38692994 +0000 UTC m=+164.905248858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.898613 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" event={"ID":"15b7bb22-0ae1-46af-a1cf-4024996c996f","Type":"ContainerStarted","Data":"b11179d0e6d69f986fde8bc8f4825704901833a3b9cf4f4b5bf9a58a587774c4"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.899632 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" event={"ID":"5b09198a-0a36-412a-86f9-f723fb6a483c","Type":"ContainerStarted","Data":"b37278012ddc49ae4f49fe38f532351716e1e28141d7cec07882c40e59a0e80c"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.900807 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" event={"ID":"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6","Type":"ContainerStarted","Data":"971c959eb07077f071dd9ba2f1923d91fff67607e1ef619a255bba9f522de24a"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.901660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" event={"ID":"d779a1a6-0749-48f9-a4a7-75e65a13c7da","Type":"ContainerStarted","Data":"f454b537ff1edd0ac501de79989b8d547d29f732838d0625984b6be68e1ac5b1"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.902901 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rdzj8" event={"ID":"ce3173af-2480-4a96-a7c5-ab9c44b18c64","Type":"ContainerStarted","Data":"9ed7cf3c85d2db4f966ce68e80da16ffb2929901542f984eff7cd9a76d2e569e"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.903005 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rdzj8" event={"ID":"ce3173af-2480-4a96-a7c5-ab9c44b18c64","Type":"ContainerStarted","Data":"6d9867c124cf8e6284c300546fc2dc7df2e530f65591a01398cc6476c54f73d4"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.903739 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" event={"ID":"2c18bce1-7e19-4745-a51d-de4a04e44dd0","Type":"ContainerStarted","Data":"db625146b1357ef1571c184d709ffb02c447d462f4570531fd93e3dbc9b48bd4"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.904683 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" event={"ID":"de33dc58-134b-4791-9b40-7529ae2c8c8b","Type":"ContainerStarted","Data":"f0591ee2b627d9c08db45e0a40e64e7e33fe551723069ab11ac81e3a6473c4d7"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.904796 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" event={"ID":"de33dc58-134b-4791-9b40-7529ae2c8c8b","Type":"ContainerStarted","Data":"6995ad8d69db3781ade84496f366f08dab35e23e74c873b7282ec1ae7b8a4504"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.905754 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" event={"ID":"33f93332-da64-4f57-b6cd-c4e643976bed","Type":"ContainerStarted","Data":"79a6dea78212c3c1760dafe2dc39769aab3427530ca2af460885712fd147a869"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.906656 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" event={"ID":"10534d55-c886-4f76-b7ba-011bee2b965d","Type":"ContainerStarted","Data":"f125eafd52fd7b0b6b1024d70c750279d8dee334d59e3db8c7fdcf38b8bb2153"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.907875 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" event={"ID":"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d","Type":"ContainerStarted","Data":"c1a85f9153efb3d4995114d173d41822f05a859346cf0d1c332a26fa9f4b0bf3"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.909018 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" event={"ID":"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c","Type":"ContainerStarted","Data":"53cfe9794a12488d2e1ec4a2ecba6cfe37cce0437c9b6db6644ad4f94b74f5d0"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.909861 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-9dtg7" event={"ID":"7aed61d7-778d-4820-881a-e087775da1a1","Type":"ContainerStarted","Data":"f4c2304e6b258c7d0749483e6ffdee1025757cb8f678a63af4b53f9851eca70c"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.910517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bvccg" event={"ID":"222d84fa-7732-4a9d-b250-e66c95092ad6","Type":"ContainerStarted","Data":"971fa1309f2fb3f0d861dd96b87f9b97c4eaf1305d13a36f5f0499dca602db76"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.911362 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" event={"ID":"96d741c5-ebc0-4c04-9192-c2661a5d0e0a","Type":"ContainerStarted","Data":"636db4924b957db389426d5bd0e19f816b10d61fe7c062020c2d57969089ae61"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.911468 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" event={"ID":"96d741c5-ebc0-4c04-9192-c2661a5d0e0a","Type":"ContainerStarted","Data":"e4f9a724d0dd5911a017fe43c95a3910d7b6daeec5c9c046eced082eb08093e0"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.912755 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbd3b91b-be30-474a-94d4-b4a684eef907" containerID="807e12af5c35f602a3c3d72fab26705503f292e6c753893e92ad325320811296" exitCode=0 Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.912884 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" event={"ID":"dbd3b91b-be30-474a-94d4-b4a684eef907","Type":"ContainerDied","Data":"807e12af5c35f602a3c3d72fab26705503f292e6c753893e92ad325320811296"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.914516 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" event={"ID":"bfdc21f9-9837-4ab1-b89f-51d630bfc302","Type":"ContainerStarted","Data":"d36c9e12721dba34ee734d7456ca9655d659ce9312ce21dc668e82aa8a964476"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.915312 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" event={"ID":"18c06562-74d0-4b4c-be8e-2fcbe69c562b","Type":"ContainerStarted","Data":"dfaf775e83e4c1347671c1c4d50a1b4e00bd12a0e46c820b67340f6d8e6b8622"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.916290 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" event={"ID":"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e","Type":"ContainerStarted","Data":"4bd0ba79a8122b8d331f4eb7350f6e423017c4c9368d048b2fd0727892140847"} Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.917628 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" event={"ID":"d8f74109-ec7a-423c-9330-8089c7935d18","Type":"ContainerStarted","Data":"208bd9eb290b9166a73dfce1ad9c6059fd0442ff2b6ee97112059e65e0a56ef4"} Nov 27 11:27:26 crc kubenswrapper[4796]: W1127 11:27:26.982215 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10b64bb3_de9d_4c66_8acc_13369ba5dc24.slice/crio-e7388ca3bd530594dd8e76239ade3fd23e6f9286e9b5918d136fd56337a1c1e0 WatchSource:0}: Error finding container e7388ca3bd530594dd8e76239ade3fd23e6f9286e9b5918d136fd56337a1c1e0: Status 404 returned error can't find the container with id e7388ca3bd530594dd8e76239ade3fd23e6f9286e9b5918d136fd56337a1c1e0 Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.991537 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.992593 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.992627 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv"] Nov 27 11:27:26 crc kubenswrapper[4796]: I1127 11:27:26.996597 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pmhfw"] Nov 27 11:27:26 crc kubenswrapper[4796]: E1127 11:27:26.997118 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.49709097 +0000 UTC m=+165.015409948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.005893 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-2vdl8"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.018946 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7t7jg"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.019014 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l954k"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.024491 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.025059 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ktgzm"] Nov 27 11:27:27 crc kubenswrapper[4796]: W1127 11:27:27.038760 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffe24f44_f2cf_4688_bb6a_dde1bba9f5a3.slice/crio-f67a29e365911ec304c75a9443118960348d519e8d3a036a8c37ecd4c0519579 WatchSource:0}: Error finding container f67a29e365911ec304c75a9443118960348d519e8d3a036a8c37ecd4c0519579: Status 404 returned error can't find the container with id f67a29e365911ec304c75a9443118960348d519e8d3a036a8c37ecd4c0519579 Nov 27 11:27:27 crc kubenswrapper[4796]: W1127 11:27:27.039059 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43316776_bbdd_4cd6_aa69_cdf9c9e5b0bb.slice/crio-348ca46461a2d7c61c4e61af447dfae016cabd8a24fa35a7d411b213f78b4458 WatchSource:0}: Error finding container 348ca46461a2d7c61c4e61af447dfae016cabd8a24fa35a7d411b213f78b4458: Status 404 returned error can't find the container with id 348ca46461a2d7c61c4e61af447dfae016cabd8a24fa35a7d411b213f78b4458 Nov 27 11:27:27 crc kubenswrapper[4796]: W1127 11:27:27.058344 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbcfbbf0_7721_4ad1_a263_e76dd5f4e8a6.slice/crio-51ce02a77679d2af33e45a4120cf9a035ffae8e58ce2754f12d9e084f9cab668 WatchSource:0}: Error finding container 51ce02a77679d2af33e45a4120cf9a035ffae8e58ce2754f12d9e084f9cab668: Status 404 returned error can't find the container with id 51ce02a77679d2af33e45a4120cf9a035ffae8e58ce2754f12d9e084f9cab668 Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.092492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.092775 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.592764897 +0000 UTC m=+165.111083815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.156576 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gqch4"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.160639 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.169200 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.169245 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 27 11:27:27 crc kubenswrapper[4796]: W1127 11:27:27.171232 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d08ea76_e06c_433f_986a_def48c08f46f.slice/crio-5f54d1ba875b5f05a1e2204381593e5bc9251bffe84eb8fd89c4dc1e311e2af3 WatchSource:0}: Error finding container 5f54d1ba875b5f05a1e2204381593e5bc9251bffe84eb8fd89c4dc1e311e2af3: Status 404 returned error can't find the container with id 5f54d1ba875b5f05a1e2204381593e5bc9251bffe84eb8fd89c4dc1e311e2af3 Nov 27 11:27:27 crc kubenswrapper[4796]: W1127 11:27:27.171931 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7abc2a6e_b355_4186_9dce_56f8ec89ef14.slice/crio-61e42da7e2cd34037f6e533cd1f026d8fedabf10c0548d451143770b2871a917 WatchSource:0}: Error finding container 61e42da7e2cd34037f6e533cd1f026d8fedabf10c0548d451143770b2871a917: Status 404 returned error can't find the container with id 61e42da7e2cd34037f6e533cd1f026d8fedabf10c0548d451143770b2871a917 Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.177499 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.193736 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.193912 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.693876244 +0000 UTC m=+165.212195162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.194089 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.194517 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.6945053 +0000 UTC m=+165.212824218 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.197126 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-v65j9" podStartSLOduration=143.197113367 podStartE2EDuration="2m23.197113367s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:27.196684797 +0000 UTC m=+164.715003725" watchObservedRunningTime="2025-11-27 11:27:27.197113367 +0000 UTC m=+164.715432275" Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.203421 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z"] Nov 27 11:27:27 crc kubenswrapper[4796]: W1127 11:27:27.224888 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod517b90c8_7ceb_43fa_9d57_a2c5ac735138.slice/crio-d9edfd04524aebc79a94e91cfa8b080dce0ebc226f1c7ff916d8fbcde7d41d0b WatchSource:0}: Error finding container d9edfd04524aebc79a94e91cfa8b080dce0ebc226f1c7ff916d8fbcde7d41d0b: Status 404 returned error can't find the container with id d9edfd04524aebc79a94e91cfa8b080dce0ebc226f1c7ff916d8fbcde7d41d0b Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.229544 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk"] Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.295247 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.295409 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.795378471 +0000 UTC m=+165.313697389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.295538 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.295948 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.795938116 +0000 UTC m=+165.314257034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.396838 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.396991 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.896966531 +0000 UTC m=+165.415285449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.397527 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.397932 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.897911455 +0000 UTC m=+165.416230383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.498833 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.498965 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:27.99893915 +0000 UTC m=+165.517258068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.504299 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.504616 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.505166 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.005125529 +0000 UTC m=+165.523444447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.507036 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0494bee3-7923-49de-8c4f-e0fa4ffad936-metrics-certs\") pod \"network-metrics-daemon-79ll4\" (UID: \"0494bee3-7923-49de-8c4f-e0fa4ffad936\") " pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.606312 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.606479 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.106455372 +0000 UTC m=+165.624774290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.606599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.606937 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.106929834 +0000 UTC m=+165.625248742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.687821 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-79ll4" Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.708988 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.709179 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.20914829 +0000 UTC m=+165.727467278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.709419 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.709854 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.209837707 +0000 UTC m=+165.728156635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.810643 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.810986 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.310949534 +0000 UTC m=+165.829268492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.911853 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:27 crc kubenswrapper[4796]: E1127 11:27:27.912563 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.412529093 +0000 UTC m=+165.930848051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.925332 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2vdl8" event={"ID":"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3","Type":"ContainerStarted","Data":"f67a29e365911ec304c75a9443118960348d519e8d3a036a8c37ecd4c0519579"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.927015 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" event={"ID":"7abc2a6e-b355-4186-9dce-56f8ec89ef14","Type":"ContainerStarted","Data":"61e42da7e2cd34037f6e533cd1f026d8fedabf10c0548d451143770b2871a917"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.931138 4796 generic.go:334] "Generic (PLEG): container finished" podID="f9d82114-9bc6-4bce-8418-46c860542aa4" containerID="87145e8071d62b790199559ff0759d49e396b998a2407901cd568b110c45e8c1" exitCode=0 Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.931499 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" event={"ID":"f9d82114-9bc6-4bce-8418-46c860542aa4","Type":"ContainerDied","Data":"87145e8071d62b790199559ff0759d49e396b998a2407901cd568b110c45e8c1"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.934851 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gqch4" event={"ID":"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365","Type":"ContainerStarted","Data":"9af273e88a887c993f97c409d2943d2b30b3a7a53a10422244c1cd03439f1caa"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.939442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-9dtg7" event={"ID":"7aed61d7-778d-4820-881a-e087775da1a1","Type":"ContainerStarted","Data":"25d64d9ede44555325dfbab5df68289d78b8f6d388c4aca0efab0838fb709dee"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.940958 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" event={"ID":"10b64bb3-de9d-4c66-8acc-13369ba5dc24","Type":"ContainerStarted","Data":"e7388ca3bd530594dd8e76239ade3fd23e6f9286e9b5918d136fd56337a1c1e0"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.942013 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" event={"ID":"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6","Type":"ContainerStarted","Data":"51ce02a77679d2af33e45a4120cf9a035ffae8e58ce2754f12d9e084f9cab668"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.943719 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" event={"ID":"fc25bf6e-f7fc-4e32-a802-ba3df37f6688","Type":"ContainerStarted","Data":"73669bc7be555b4d2cd7b244c1d6b66fcb68b5e3a914fdab8ce36a824d81a673"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.945118 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerStarted","Data":"93b39b7b5f804c3efd2862116a665de39994507683dcf1ad9fa62081ddb3da51"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.946031 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" event={"ID":"95abf6de-142f-47bc-9492-38fc1e8f69b0","Type":"ContainerStarted","Data":"fbc426a86fb1c68398fcbe1a7c33393eb7fab1f4eec3316bf362c297c03ecb58"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.946980 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" event={"ID":"7aa958d0-0f0b-46c2-884f-e7f23431c022","Type":"ContainerStarted","Data":"9013135d2607b29eb779e7575ab55a97f90ea835d8399a563f52aa1752dc3a33"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.949232 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" event={"ID":"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6","Type":"ContainerStarted","Data":"9df4e80a5deb85d98ac5855c841723b8731fc275c633931826c85a8444f93d9c"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.952980 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" event={"ID":"0a40ce6c-6834-40bd-aad5-a34773f4e91b","Type":"ContainerStarted","Data":"50039488a78daa26567bcac7d2262572f7d6e76f8cde0a76acdf89f281965efc"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.957405 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" event={"ID":"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e","Type":"ContainerStarted","Data":"8d95a78192f9ce44682a731505df629b8b42cf37b1ecf9732745553f5c00338c"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.962385 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" event={"ID":"0d08ea76-e06c-433f-986a-def48c08f46f","Type":"ContainerStarted","Data":"5f54d1ba875b5f05a1e2204381593e5bc9251bffe84eb8fd89c4dc1e311e2af3"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.965490 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" event={"ID":"517b90c8-7ceb-43fa-9d57-a2c5ac735138","Type":"ContainerStarted","Data":"d9edfd04524aebc79a94e91cfa8b080dce0ebc226f1c7ff916d8fbcde7d41d0b"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.967486 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" event={"ID":"3e6acd38-f2b9-4628-8d87-c269ef09d9bc","Type":"ContainerStarted","Data":"4111275a7e734d16d9ae230c4e543aa7744272517eeac4baa3f42641115d511c"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.968680 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" event={"ID":"5f7f0a02-e94f-409c-8ed0-c5a7b301bace","Type":"ContainerStarted","Data":"82dab821ffc404e5e28f8b9e4f4c3abc99b43fcd1475956690c97ea4ae39ef96"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.970019 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" event={"ID":"15b7bb22-0ae1-46af-a1cf-4024996c996f","Type":"ContainerStarted","Data":"ef143237add4067b0cd364d2fade2c03c01c827b227cf3ef098e57ad05525b22"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.971100 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" event={"ID":"aba9e15a-aeb7-40f8-8c19-e68305bf379e","Type":"ContainerStarted","Data":"b519d072b6e9806626896ae5c8f3fa9116714be2b7272a659bab67ec4b676d70"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.972360 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" event={"ID":"d779a1a6-0749-48f9-a4a7-75e65a13c7da","Type":"ContainerStarted","Data":"8e827ad48da4048f8d54dbfddc3aea380f889444c9ace8ad9a15aa3baadb89e4"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.973476 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" event={"ID":"f8096c4b-d1ba-471a-9d05-22aab449b51c","Type":"ContainerStarted","Data":"06d45d3b45eceaf1d85449e22023120ecdaddf6f9041c9377bd91d4e9d8b7479"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.974353 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" event={"ID":"b4726245-f4f8-494f-b2d9-2d8cb83f05e3","Type":"ContainerStarted","Data":"93d1df29f58abbf1eede9c3aa0647a6295cd3f4ed667f42a74a552caf1f3f47e"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.975171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l954k" event={"ID":"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb","Type":"ContainerStarted","Data":"348ca46461a2d7c61c4e61af447dfae016cabd8a24fa35a7d411b213f78b4458"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.976038 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" event={"ID":"9297600d-7a17-4a87-9ed1-26357ea3df51","Type":"ContainerStarted","Data":"5567de5776d2865d2800413a5726681f9f30067285d4e145d19e64877c0d6f20"} Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.976478 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.980474 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:27 crc kubenswrapper[4796]: I1127 11:27:27.980564 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.013015 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.013156 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.513129317 +0000 UTC m=+166.031448245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.014661 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.015491 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.515452197 +0000 UTC m=+166.033771155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.117705 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.119580 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.619563871 +0000 UTC m=+166.137882789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.165749 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.166220 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.199819 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-skdgb" podStartSLOduration=145.199802383 podStartE2EDuration="2m25.199802383s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:28.193710576 +0000 UTC m=+165.712029494" watchObservedRunningTime="2025-11-27 11:27:28.199802383 +0000 UTC m=+165.718121301" Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.236598 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.236879 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.736868254 +0000 UTC m=+166.255187172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.253527 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rdzj8" podStartSLOduration=144.253509962 podStartE2EDuration="2m24.253509962s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:28.251698106 +0000 UTC m=+165.770017024" watchObservedRunningTime="2025-11-27 11:27:28.253509962 +0000 UTC m=+165.771828880" Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.337479 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.337632 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.837607382 +0000 UTC m=+166.355926300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.338015 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.338389 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.838374111 +0000 UTC m=+166.356693029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.439284 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.439608 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:28.939595592 +0000 UTC m=+166.457914510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.541144 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.541584 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.041553711 +0000 UTC m=+166.559872629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.644374 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.645018 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.145003228 +0000 UTC m=+166.663322146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.745712 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.746847 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.246835243 +0000 UTC m=+166.765154161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.848077 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.848230 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.348201877 +0000 UTC m=+166.866520815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.848925 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.849240 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.349228954 +0000 UTC m=+166.867547872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.867584 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-79ll4"] Nov 27 11:27:28 crc kubenswrapper[4796]: W1127 11:27:28.880980 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0494bee3_7923_49de_8c4f_e0fa4ffad936.slice/crio-db46f6c59d7faf91a90b091d32b8b51abac6f22bc36bdbfddf5b05cf809a033a WatchSource:0}: Error finding container db46f6c59d7faf91a90b091d32b8b51abac6f22bc36bdbfddf5b05cf809a033a: Status 404 returned error can't find the container with id db46f6c59d7faf91a90b091d32b8b51abac6f22bc36bdbfddf5b05cf809a033a Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.949711 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:28 crc kubenswrapper[4796]: E1127 11:27:28.950146 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.450130596 +0000 UTC m=+166.968449514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.990488 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" event={"ID":"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6","Type":"ContainerStarted","Data":"2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927"} Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.990777 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.992419 4796 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-l8bvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.992463 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 27 11:27:28 crc kubenswrapper[4796]: I1127 11:27:28.998902 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" event={"ID":"e7f8ca38-34d4-4e26-b1e5-8798f9986d4c","Type":"ContainerStarted","Data":"2911b088ea36da9baa98a5c3a9d26f7397590a8573786f5f83b9fb77b5f975a2"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.002278 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" event={"ID":"5b09198a-0a36-412a-86f9-f723fb6a483c","Type":"ContainerStarted","Data":"04115171e01ae9b5f18873030202069799fdf0cde426fa9ee19605addba1bd31"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.007402 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" podStartSLOduration=145.007382956 podStartE2EDuration="2m25.007382956s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.005457177 +0000 UTC m=+166.523776105" watchObservedRunningTime="2025-11-27 11:27:29.007382956 +0000 UTC m=+166.525701874" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.010403 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" event={"ID":"10534d55-c886-4f76-b7ba-011bee2b965d","Type":"ContainerStarted","Data":"5ad4b6e2e5e5bd271da41f5df2ed478a4515a76f8d148429acfca626ca891530"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.012414 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" event={"ID":"bfdc21f9-9837-4ab1-b89f-51d630bfc302","Type":"ContainerStarted","Data":"bd818513b02c5d2c30b2f4e5cfe2bcd80fe658f12e3c2749e788d497cd59f55b"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.022743 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" event={"ID":"33f93332-da64-4f57-b6cd-c4e643976bed","Type":"ContainerStarted","Data":"1e03f4fabe9c098d1ec1adf7ff29128e67947b38a61b3d92893468b3fc990753"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.030788 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kvksm" podStartSLOduration=145.030770447 podStartE2EDuration="2m25.030770447s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.030022387 +0000 UTC m=+166.548341325" watchObservedRunningTime="2025-11-27 11:27:29.030770447 +0000 UTC m=+166.549089365" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.047853 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bvccg" event={"ID":"222d84fa-7732-4a9d-b250-e66c95092ad6","Type":"ContainerStarted","Data":"f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.049392 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" event={"ID":"dbd3b91b-be30-474a-94d4-b4a684eef907","Type":"ContainerStarted","Data":"bef58accb2eaaa893b7f8e4130370567c2a3f9ef95f1933cf0456d2aea65450c"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.055326 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.055991 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.555976335 +0000 UTC m=+167.074295263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.057405 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8tdzt" podStartSLOduration=145.05739469 podStartE2EDuration="2m25.05739469s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.055186003 +0000 UTC m=+166.573504921" watchObservedRunningTime="2025-11-27 11:27:29.05739469 +0000 UTC m=+166.575713608" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.077095 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" event={"ID":"18c06562-74d0-4b4c-be8e-2fcbe69c562b","Type":"ContainerStarted","Data":"b484f30d7cc9d0b2c6e8813ecaf0dd7b64919dd1c8966f2a154328ffaca6266e"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.078836 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4lbbd" podStartSLOduration=145.078818301 podStartE2EDuration="2m25.078818301s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.076894721 +0000 UTC m=+166.595213639" watchObservedRunningTime="2025-11-27 11:27:29.078818301 +0000 UTC m=+166.597137219" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.092540 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" event={"ID":"d8f74109-ec7a-423c-9330-8089c7935d18","Type":"ContainerStarted","Data":"81d9078b60dfb50d53ec4246749d035b14db8c468b46d1480c4dfb13fa59bf59"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.094519 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-79ll4" event={"ID":"0494bee3-7923-49de-8c4f-e0fa4ffad936","Type":"ContainerStarted","Data":"db46f6c59d7faf91a90b091d32b8b51abac6f22bc36bdbfddf5b05cf809a033a"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.110390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" event={"ID":"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d","Type":"ContainerStarted","Data":"75b6a34164bb041dacdb1f2700979a30b2b13159bb6a415cf5ec67ddc67f634c"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.117708 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" event={"ID":"2c18bce1-7e19-4745-a51d-de4a04e44dd0","Type":"ContainerStarted","Data":"cf54f9763fc6ff5f87c050e275d2e05de3343b6aca915327aa278e2328058b95"} Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.118198 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mz8zx" podStartSLOduration=145.118186572 podStartE2EDuration="2m25.118186572s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.117476874 +0000 UTC m=+166.635795792" watchObservedRunningTime="2025-11-27 11:27:29.118186572 +0000 UTC m=+166.636505480" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.118249 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.118301 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.149895 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-z8hd2" podStartSLOduration=145.149875156 podStartE2EDuration="2m25.149875156s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.147289139 +0000 UTC m=+166.665608057" watchObservedRunningTime="2025-11-27 11:27:29.149875156 +0000 UTC m=+166.668194074" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.156634 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.156793 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.656768183 +0000 UTC m=+167.175087101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.157092 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.158624 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.65861672 +0000 UTC m=+167.176935628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.163019 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8wmkg" podStartSLOduration=145.162998983 podStartE2EDuration="2m25.162998983s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.16208661 +0000 UTC m=+166.680405528" watchObservedRunningTime="2025-11-27 11:27:29.162998983 +0000 UTC m=+166.681317911" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.164294 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.164342 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.258482 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.260544 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.760523048 +0000 UTC m=+167.278841966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.360534 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.360901 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.860884515 +0000 UTC m=+167.379203433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.461125 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.461375 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.961344676 +0000 UTC m=+167.479663594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.461499 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.461869 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:29.96185805 +0000 UTC m=+167.480176968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.562955 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.563157 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.063128231 +0000 UTC m=+167.581447159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.563454 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.563912 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.06388979 +0000 UTC m=+167.582208758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.664472 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.664675 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.164636018 +0000 UTC m=+167.682954966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.664841 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.665198 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.165182652 +0000 UTC m=+167.683501580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.766038 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.766231 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.266195917 +0000 UTC m=+167.784514845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.766315 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.766633 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.266621787 +0000 UTC m=+167.784940705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.867757 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.868652 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.368600757 +0000 UTC m=+167.886919675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:29 crc kubenswrapper[4796]: I1127 11:27:29.969759 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:29 crc kubenswrapper[4796]: E1127 11:27:29.970114 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.470099254 +0000 UTC m=+167.988418172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.071189 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.071594 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.57155755 +0000 UTC m=+168.089876498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.126573 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" event={"ID":"517b90c8-7ceb-43fa-9d57-a2c5ac735138","Type":"ContainerStarted","Data":"ca7cdff32f53613f1dcc6af79d4259fd715a18e96f14b318170bd49c641ef783"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.128422 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" event={"ID":"96d741c5-ebc0-4c04-9192-c2661a5d0e0a","Type":"ContainerStarted","Data":"14dc195eb84527eb04cf3d3d45530a7f98caa36e61f2b9c2620e1e9631847b5f"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.130589 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" event={"ID":"8bc8576c-6bf0-4215-8dfb-6d0320d9eb8e","Type":"ContainerStarted","Data":"cd05348ccd0798258063560ee283a52aca6fe56e38aef0a87a779ef99c3b005b"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.132433 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" event={"ID":"7abc2a6e-b355-4186-9dce-56f8ec89ef14","Type":"ContainerStarted","Data":"2385be8e33fdc39127ebf08281a0a9756885218852d33b6fe6e5dc3791fe7c69"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.134126 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" event={"ID":"b4726245-f4f8-494f-b2d9-2d8cb83f05e3","Type":"ContainerStarted","Data":"a31b19d1c008a8b977505cf8d6b545c8492a2947c14cdfa6930be7e4a5339a7a"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.135992 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l954k" event={"ID":"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb","Type":"ContainerStarted","Data":"8e72e23d08543cbee781e18034559a5bd5dc64369d3b8d21f732c1a358028094"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.137774 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" event={"ID":"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6","Type":"ContainerStarted","Data":"b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.139962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" event={"ID":"10b64bb3-de9d-4c66-8acc-13369ba5dc24","Type":"ContainerStarted","Data":"c8b9c51353d6dad8d4feaee2c9a3ffa727c700915bda5778de66369f36c76e4e"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.141677 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2vdl8" event={"ID":"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3","Type":"ContainerStarted","Data":"11efd3c2372eb6fa82c8d1cb1243c6db70173c9aea204622595a20204852530e"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.144231 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" event={"ID":"cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6","Type":"ContainerStarted","Data":"7b183f06969762f6d08c4df43ca7cf5c310b9cd8d8d33f58f971c883c7d5bee0"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.146072 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" event={"ID":"0a40ce6c-6834-40bd-aad5-a34773f4e91b","Type":"ContainerStarted","Data":"63aa7d68e3eaa2297b89cb4c2e688f72417e75929ef94d7aa849e5d71697ba3d"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.148580 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" event={"ID":"de33dc58-134b-4791-9b40-7529ae2c8c8b","Type":"ContainerStarted","Data":"93419a06c1b664d896b197dd5cce52fa5820220b0627eea8a0e99817b9bac7d2"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.150930 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" event={"ID":"f8096c4b-d1ba-471a-9d05-22aab449b51c","Type":"ContainerStarted","Data":"b930bf5056d5e639f528c3785d2391620d153941342cb357908cb885e7186fa9"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.153817 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" event={"ID":"f9d82114-9bc6-4bce-8418-46c860542aa4","Type":"ContainerStarted","Data":"a5630410b12dc71a2bd859091e1a513f52034e31988afc6bf63c491c2fe35f2e"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.155868 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" event={"ID":"3e6acd38-f2b9-4628-8d87-c269ef09d9bc","Type":"ContainerStarted","Data":"41a0629854a630c4ead2bddff82b15d85609f0abbdd4766d4be86138d44857ac"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.157565 4796 generic.go:334] "Generic (PLEG): container finished" podID="95abf6de-142f-47bc-9492-38fc1e8f69b0" containerID="d9f52519a396371f0d65fe75e71cc831f983fc305ceefc4cb7a4a2cef6931782" exitCode=0 Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.157622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" event={"ID":"95abf6de-142f-47bc-9492-38fc1e8f69b0","Type":"ContainerDied","Data":"d9f52519a396371f0d65fe75e71cc831f983fc305ceefc4cb7a4a2cef6931782"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.159968 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" event={"ID":"9297600d-7a17-4a87-9ed1-26357ea3df51","Type":"ContainerStarted","Data":"299c7052f7d6546f9849981bf98b33114f042c5ea14a12fc07965930b9c95de6"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.162534 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gqch4" event={"ID":"39c826c0-9aa1-4fed-a8ca-3dfafe5f5365","Type":"ContainerStarted","Data":"e17f72c6af39b4921ed9db000a729d0bda3b7f09e54cea228ed267bc74ce4112"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.164059 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" event={"ID":"7aa958d0-0f0b-46c2-884f-e7f23431c022","Type":"ContainerStarted","Data":"8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.165745 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:30 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:30 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:30 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.165845 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.166331 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" event={"ID":"aba9e15a-aeb7-40f8-8c19-e68305bf379e","Type":"ContainerStarted","Data":"1f4a3593d20958f1a824164c9d481778aefe9d9c4756421c92462fbac9201729"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.167975 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" event={"ID":"fc25bf6e-f7fc-4e32-a802-ba3df37f6688","Type":"ContainerStarted","Data":"e4a1ab431f3e3afc6e1af1b98ba5c15242e4976fef2d1211fb1944ceec7820eb"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.169471 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" event={"ID":"0d08ea76-e06c-433f-986a-def48c08f46f","Type":"ContainerStarted","Data":"1833bdbb8a153f2e9257cf62feb39e3b7abcc43c94ef51d60f2ab48c4703a657"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.171989 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerStarted","Data":"eab8c03db19581f6cd9186955c3d21d66771e944336a1df0454ef7c0a4e6d96e"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.172235 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.172617 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.672601805 +0000 UTC m=+168.190920723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.177116 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" event={"ID":"0d8269c0-b923-4ad0-8aa8-ddb2e21b257d","Type":"ContainerStarted","Data":"e5039f76040dfcf9b596f8adbebda9ee17441f7e80fdf3e4ae757908d41a1518"} Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.179345 4796 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-l8bvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.179413 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.179518 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.195748 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-bvccg" podStartSLOduration=146.195721859 podStartE2EDuration="2m26.195721859s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:30.193884012 +0000 UTC m=+167.712202930" watchObservedRunningTime="2025-11-27 11:27:30.195721859 +0000 UTC m=+167.714040807" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.197536 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-9dtg7" podStartSLOduration=7.197526806 podStartE2EDuration="7.197526806s" podCreationTimestamp="2025-11-27 11:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:29.18041468 +0000 UTC m=+166.698733598" watchObservedRunningTime="2025-11-27 11:27:30.197526806 +0000 UTC m=+167.715845754" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.214595 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lsx6g" podStartSLOduration=147.214570383 podStartE2EDuration="2m27.214570383s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:30.209278787 +0000 UTC m=+167.727597705" watchObservedRunningTime="2025-11-27 11:27:30.214570383 +0000 UTC m=+167.732889331" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.226852 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" podStartSLOduration=147.226836319 podStartE2EDuration="2m27.226836319s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:30.226007387 +0000 UTC m=+167.744326305" watchObservedRunningTime="2025-11-27 11:27:30.226836319 +0000 UTC m=+167.745155237" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.241287 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8qplk" podStartSLOduration=147.241250799 podStartE2EDuration="2m27.241250799s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:30.240048048 +0000 UTC m=+167.758366966" watchObservedRunningTime="2025-11-27 11:27:30.241250799 +0000 UTC m=+167.759569717" Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.273548 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.273729 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.773704272 +0000 UTC m=+168.292023180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.274043 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.276588 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.776572846 +0000 UTC m=+168.294891764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.375682 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.376201 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.876178275 +0000 UTC m=+168.394497213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.477788 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.478191 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:30.978170754 +0000 UTC m=+168.496489692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.579085 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.579291 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.07923562 +0000 UTC m=+168.597554548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.579426 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.579831 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.079821366 +0000 UTC m=+168.598140354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.680571 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.680767 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.180734928 +0000 UTC m=+168.699053846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.681008 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.681342 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.181329113 +0000 UTC m=+168.699648031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.781849 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.782066 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.28203054 +0000 UTC m=+168.800349458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.782199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.782580 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.282562263 +0000 UTC m=+168.800881191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.883026 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.883202 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.383166937 +0000 UTC m=+168.901485895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.883317 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.883798 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.383782243 +0000 UTC m=+168.902101191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.984727 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.984892 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.484872729 +0000 UTC m=+169.003191647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:30 crc kubenswrapper[4796]: I1127 11:27:30.984966 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:30 crc kubenswrapper[4796]: E1127 11:27:30.985349 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.485336131 +0000 UTC m=+169.003655049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.085860 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.086015 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.585992257 +0000 UTC m=+169.104311185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.086139 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.086486 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.586474739 +0000 UTC m=+169.104793657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.172167 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:31 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:31 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:31 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.172248 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.182776 4796 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zs7tf container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.182825 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" podUID="dbd3b91b-be30-474a-94d4-b4a684eef907" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.182842 4796 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zs7tf container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.182866 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" podUID="dbd3b91b-be30-474a-94d4-b4a684eef907" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.185791 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-79ll4" event={"ID":"0494bee3-7923-49de-8c4f-e0fa4ffad936","Type":"ContainerStarted","Data":"0a55d13e5c2a6ccac12a034fef21e73d8595110dcb9a80437c5690d408fdd8ff"} Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.191973 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192005 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192015 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192025 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192034 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192043 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192040 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192008 4796 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-g949p container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192059 4796 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sngfr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192079 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" podUID="fc25bf6e-f7fc-4e32-a802-ba3df37f6688" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192197 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.192219 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.692190085 +0000 UTC m=+169.210509013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192039 4796 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cwptf container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192052 4796 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zs7tf container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192253 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" podUID="3e6acd38-f2b9-4628-8d87-c269ef09d9bc" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192276 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" podUID="dbd3b91b-be30-474a-94d4-b4a684eef907" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192107 4796 patch_prober.go:28] interesting pod/console-operator-58897d9998-zbnzg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192302 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" podUID="10b64bb3-de9d-4c66-8acc-13369ba5dc24" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192258 4796 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6nhkv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192385 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" podUID="cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.192807 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.193117 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.693105939 +0000 UTC m=+169.211424857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.193473 4796 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fdwzj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.193861 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.210182 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" podStartSLOduration=147.210157767 podStartE2EDuration="2m27.210157767s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.205764813 +0000 UTC m=+168.724083741" watchObservedRunningTime="2025-11-27 11:27:31.210157767 +0000 UTC m=+168.728476685" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.246428 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" podStartSLOduration=147.246406698 podStartE2EDuration="2m27.246406698s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.244055737 +0000 UTC m=+168.762374675" watchObservedRunningTime="2025-11-27 11:27:31.246406698 +0000 UTC m=+168.764725626" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.272940 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-bxf5l" podStartSLOduration=147.272919428 podStartE2EDuration="2m27.272919428s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.272287362 +0000 UTC m=+168.790606290" watchObservedRunningTime="2025-11-27 11:27:31.272919428 +0000 UTC m=+168.791238356" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.293714 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.293910 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.793878047 +0000 UTC m=+169.312196975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.294194 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.304611 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.804592221 +0000 UTC m=+169.322911139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.310871 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" podStartSLOduration=147.310845763 podStartE2EDuration="2m27.310845763s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.296754241 +0000 UTC m=+168.815073159" watchObservedRunningTime="2025-11-27 11:27:31.310845763 +0000 UTC m=+168.829164681" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.334485 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-gqch4" podStartSLOduration=9.334464609 podStartE2EDuration="9.334464609s" podCreationTimestamp="2025-11-27 11:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.327809409 +0000 UTC m=+168.846128327" watchObservedRunningTime="2025-11-27 11:27:31.334464609 +0000 UTC m=+168.852783527" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.348576 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" podStartSLOduration=147.348556491 podStartE2EDuration="2m27.348556491s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.347817832 +0000 UTC m=+168.866136750" watchObservedRunningTime="2025-11-27 11:27:31.348556491 +0000 UTC m=+168.866875409" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.365706 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7rks7" podStartSLOduration=147.365682141 podStartE2EDuration="2m27.365682141s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.364734197 +0000 UTC m=+168.883053115" watchObservedRunningTime="2025-11-27 11:27:31.365682141 +0000 UTC m=+168.884001079" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.396528 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.396888 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.896869652 +0000 UTC m=+169.415188570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.408791 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" podStartSLOduration=147.408771738 podStartE2EDuration="2m27.408771738s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.40849554 +0000 UTC m=+168.926814498" watchObservedRunningTime="2025-11-27 11:27:31.408771738 +0000 UTC m=+168.927090656" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.410627 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" podStartSLOduration=148.410619785 podStartE2EDuration="2m28.410619785s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.389047841 +0000 UTC m=+168.907366749" watchObservedRunningTime="2025-11-27 11:27:31.410619785 +0000 UTC m=+168.928938703" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.421166 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-x97jq" podStartSLOduration=147.421148216 podStartE2EDuration="2m27.421148216s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.420799467 +0000 UTC m=+168.939118385" watchObservedRunningTime="2025-11-27 11:27:31.421148216 +0000 UTC m=+168.939467134" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.439254 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-85t6r" podStartSLOduration=147.43923671 podStartE2EDuration="2m27.43923671s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:31.436594132 +0000 UTC m=+168.954913050" watchObservedRunningTime="2025-11-27 11:27:31.43923671 +0000 UTC m=+168.957555628" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.498074 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.498953 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:31.998936024 +0000 UTC m=+169.517255002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.601485 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.601722 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.101691193 +0000 UTC m=+169.620010121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.603055 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.603501 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.103463268 +0000 UTC m=+169.621782266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.704061 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.704582 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.204554405 +0000 UTC m=+169.722873393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.810614 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.810974 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.310961068 +0000 UTC m=+169.829279986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.879374 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.879424 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.911696 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.911890 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.41185918 +0000 UTC m=+169.930178098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:31 crc kubenswrapper[4796]: I1127 11:27:31.912030 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:31 crc kubenswrapper[4796]: E1127 11:27:31.912384 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.412373393 +0000 UTC m=+169.930692311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.013734 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.013917 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.513887821 +0000 UTC m=+170.032206739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.014324 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.014656 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.51464164 +0000 UTC m=+170.032960558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.115395 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.115559 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.615525242 +0000 UTC m=+170.133844160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.115779 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.116094 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.616086766 +0000 UTC m=+170.134405684 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.164853 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:32 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:32 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:32 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.164910 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.191224 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-79ll4" event={"ID":"0494bee3-7923-49de-8c4f-e0fa4ffad936","Type":"ContainerStarted","Data":"d93f221afd51ce74affa0773fe31afd2f76000015351b1fa009203ab35891e9d"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.192831 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2vdl8" event={"ID":"ffe24f44-f2cf-4688-bb6a-dde1bba9f5a3","Type":"ContainerStarted","Data":"3bfd8ac8f55335bfb4376bef2d3e526adc4950578a089ba6e8595929ac27aba9"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.192952 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.194122 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" event={"ID":"0a40ce6c-6834-40bd-aad5-a34773f4e91b","Type":"ContainerStarted","Data":"407828a28030658b1dfc10ac9a3cf15de71bbb28b530f5b5d417eb379a00a837"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.195578 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" event={"ID":"0d08ea76-e06c-433f-986a-def48c08f46f","Type":"ContainerStarted","Data":"2ef166df2ef360bcc2d778b5ec4379e0838f2c1f29333aa68c1f5182c10b9fbd"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.197143 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" event={"ID":"18c06562-74d0-4b4c-be8e-2fcbe69c562b","Type":"ContainerStarted","Data":"9edc377468acdb376c39d176b44f3e56b44c172bcd0a0ba7eeae7856b813f691"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.198942 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" event={"ID":"95abf6de-142f-47bc-9492-38fc1e8f69b0","Type":"ContainerStarted","Data":"9daa95b4ef274345fb16c57124eda8e6331e7c6d39f9fc3845630994296b9105"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.200565 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l954k" event={"ID":"43316776-bbdd-4cd6-aa69-cdf9c9e5b0bb","Type":"ContainerStarted","Data":"8e7fb187419037a1bbf971a0a350735fa7d85d7d60b70a80c6caf43009145cc7"} Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.201567 4796 patch_prober.go:28] interesting pod/console-operator-58897d9998-zbnzg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.201593 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" podUID="10b64bb3-de9d-4c66-8acc-13369ba5dc24" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.201842 4796 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-sngfr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.201864 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202239 4796 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6nhkv container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202279 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" podUID="cbcfbbf0-7721-4ad1-a263-e76dd5f4e8a6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202344 4796 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fdwzj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202392 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202522 4796 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-g949p container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202553 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" podUID="fc25bf6e-f7fc-4e32-a802-ba3df37f6688" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202751 4796 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cwptf container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.202776 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" podUID="3e6acd38-f2b9-4628-8d87-c269ef09d9bc" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.210456 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-79ll4" podStartSLOduration=148.21043878 podStartE2EDuration="2m28.21043878s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.20771781 +0000 UTC m=+169.726036728" watchObservedRunningTime="2025-11-27 11:27:32.21043878 +0000 UTC m=+169.728757698" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.216247 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.216366 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.716347631 +0000 UTC m=+170.234666549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.216552 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.216864 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.716856604 +0000 UTC m=+170.235175522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.222545 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4kcl7" podStartSLOduration=148.22252887 podStartE2EDuration="2m28.22252887s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.22058591 +0000 UTC m=+169.738904828" watchObservedRunningTime="2025-11-27 11:27:32.22252887 +0000 UTC m=+169.740847788" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.254307 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gd7r5" podStartSLOduration=148.254285485 podStartE2EDuration="2m28.254285485s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.234906707 +0000 UTC m=+169.753225625" watchObservedRunningTime="2025-11-27 11:27:32.254285485 +0000 UTC m=+169.772604403" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.258104 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-2vdl8" podStartSLOduration=9.258089594 podStartE2EDuration="9.258089594s" podCreationTimestamp="2025-11-27 11:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.257377465 +0000 UTC m=+169.775696383" watchObservedRunningTime="2025-11-27 11:27:32.258089594 +0000 UTC m=+169.776408512" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.288196 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" podStartSLOduration=148.288177546 podStartE2EDuration="2m28.288177546s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.2770489 +0000 UTC m=+169.795367818" watchObservedRunningTime="2025-11-27 11:27:32.288177546 +0000 UTC m=+169.806496464" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.305318 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4fq6z" podStartSLOduration=148.305301736 podStartE2EDuration="2m28.305301736s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.28989823 +0000 UTC m=+169.808217148" watchObservedRunningTime="2025-11-27 11:27:32.305301736 +0000 UTC m=+169.823620654" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.317356 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.319344 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.819323837 +0000 UTC m=+170.337642765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.320529 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fgr9n" podStartSLOduration=149.320512907 podStartE2EDuration="2m29.320512907s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.306027705 +0000 UTC m=+169.824346613" watchObservedRunningTime="2025-11-27 11:27:32.320512907 +0000 UTC m=+169.838831825" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.339186 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" podStartSLOduration=149.339166736 podStartE2EDuration="2m29.339166736s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.321078961 +0000 UTC m=+169.839397879" watchObservedRunningTime="2025-11-27 11:27:32.339166736 +0000 UTC m=+169.857485654" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.362088 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lxb9g" podStartSLOduration=148.362073514 podStartE2EDuration="2m28.362073514s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.346450633 +0000 UTC m=+169.864769551" watchObservedRunningTime="2025-11-27 11:27:32.362073514 +0000 UTC m=+169.880392432" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.365024 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" podStartSLOduration=148.36501503 podStartE2EDuration="2m28.36501503s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.359449487 +0000 UTC m=+169.877768405" watchObservedRunningTime="2025-11-27 11:27:32.36501503 +0000 UTC m=+169.883333948" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.387216 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2bbpf" podStartSLOduration=148.38720149 podStartE2EDuration="2m28.38720149s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.386859951 +0000 UTC m=+169.905178889" watchObservedRunningTime="2025-11-27 11:27:32.38720149 +0000 UTC m=+169.905520408" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.419720 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ktgzm" podStartSLOduration=148.419702114 podStartE2EDuration="2m28.419702114s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:32.416559863 +0000 UTC m=+169.934878781" watchObservedRunningTime="2025-11-27 11:27:32.419702114 +0000 UTC m=+169.938021032" Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.420165 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.420533 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:32.920520656 +0000 UTC m=+170.438839574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.521257 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.521406 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.021386267 +0000 UTC m=+170.539705195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.521686 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.521926 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.02191888 +0000 UTC m=+170.540237798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.622720 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.623180 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.12316081 +0000 UTC m=+170.641479738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.724216 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.724503 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.224491463 +0000 UTC m=+170.742810381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.824695 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.824875 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.324841421 +0000 UTC m=+170.843160329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.825078 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.825404 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.325396235 +0000 UTC m=+170.843715153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.926105 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.926296 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.426250365 +0000 UTC m=+170.944569293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:32 crc kubenswrapper[4796]: I1127 11:27:32.926440 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:32 crc kubenswrapper[4796]: E1127 11:27:32.926799 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.426787539 +0000 UTC m=+170.945106457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.028091 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.028309 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.528277767 +0000 UTC m=+171.046596675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.028478 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.028848 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.528838331 +0000 UTC m=+171.047157249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.129463 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.129609 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.629592159 +0000 UTC m=+171.147911077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.129704 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.129970 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.629963778 +0000 UTC m=+171.148282696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.165740 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:33 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:33 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:33 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.165831 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.230440 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" podStartSLOduration=149.230427869 podStartE2EDuration="2m29.230427869s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:33.22347952 +0000 UTC m=+170.741798438" watchObservedRunningTime="2025-11-27 11:27:33.230427869 +0000 UTC m=+170.748746777" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.230800 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.230991 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.730962932 +0000 UTC m=+171.249281850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.231137 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.231493 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.731480485 +0000 UTC m=+171.249799403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.243684 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-l954k" podStartSLOduration=149.243664019 podStartE2EDuration="2m29.243664019s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:33.241133194 +0000 UTC m=+170.759452112" watchObservedRunningTime="2025-11-27 11:27:33.243664019 +0000 UTC m=+170.761982937" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.332045 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.332198 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.832176032 +0000 UTC m=+171.350494950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.332490 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.333497 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.833483157 +0000 UTC m=+171.351802075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.381685 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.382339 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.383977 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.385502 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.390279 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.433953 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.434126 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.93410294 +0000 UTC m=+171.452421858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.434359 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.434657 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:33.934648875 +0000 UTC m=+171.452967783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.535227 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.535451 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.035412883 +0000 UTC m=+171.553731841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.535528 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71c9c343-afe0-4330-a97c-5600a14d569e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.535743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.535935 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71c9c343-afe0-4330-a97c-5600a14d569e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.536062 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.036050519 +0000 UTC m=+171.554369437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.637154 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.637357 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.137319451 +0000 UTC m=+171.655638399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.637428 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71c9c343-afe0-4330-a97c-5600a14d569e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.637548 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.637582 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71c9c343-afe0-4330-a97c-5600a14d569e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.637647 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71c9c343-afe0-4330-a97c-5600a14d569e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.637907 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.137895275 +0000 UTC m=+171.656214203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.684181 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71c9c343-afe0-4330-a97c-5600a14d569e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.711907 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.738876 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.739171 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.239148306 +0000 UTC m=+171.757467224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.846089 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.846552 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.346536825 +0000 UTC m=+171.864855743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.921071 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.937434 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 11:27:33 crc kubenswrapper[4796]: I1127 11:27:33.950642 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:33 crc kubenswrapper[4796]: E1127 11:27:33.951336 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.451318326 +0000 UTC m=+171.969637244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.051595 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.051996 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.551977442 +0000 UTC m=+172.070296360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.158054 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.159352 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.659314288 +0000 UTC m=+172.177633206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.159420 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.159809 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.659794791 +0000 UTC m=+172.178113709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.167422 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:34 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:34 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:34 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.167493 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.178104 4796 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zs7tf container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.178172 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" podUID="dbd3b91b-be30-474a-94d4-b4a684eef907" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.180070 4796 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zs7tf container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.180120 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" podUID="dbd3b91b-be30-474a-94d4-b4a684eef907" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.227247 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71c9c343-afe0-4330-a97c-5600a14d569e","Type":"ContainerStarted","Data":"98cfda4815d863acaa59c93e3618c212af9bfd8d888a5843e2489719b9592cbd"} Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.229108 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" event={"ID":"95abf6de-142f-47bc-9492-38fc1e8f69b0","Type":"ContainerStarted","Data":"7ac71c56164d1e53506c3460836f7d50b414a3c14270f699794de03202b42304"} Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.249089 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" podStartSLOduration=151.249070804 podStartE2EDuration="2m31.249070804s" podCreationTimestamp="2025-11-27 11:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:34.246474107 +0000 UTC m=+171.764793045" watchObservedRunningTime="2025-11-27 11:27:34.249070804 +0000 UTC m=+171.767389722" Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.261057 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.261906 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.761883093 +0000 UTC m=+172.280202011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.362749 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.363761 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.863744899 +0000 UTC m=+172.382063927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.463391 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.463512 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.963494142 +0000 UTC m=+172.481813060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.463650 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.463915 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:34.963908382 +0000 UTC m=+172.482227300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.564897 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.565042 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.065019919 +0000 UTC m=+172.583338837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.565613 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.565904 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.065897363 +0000 UTC m=+172.584216281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.666451 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.666628 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.166602808 +0000 UTC m=+172.684921726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.666671 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.666987 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.166973419 +0000 UTC m=+172.685292337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.767313 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.767504 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.26748132 +0000 UTC m=+172.785800238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.868284 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.868612 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.368595218 +0000 UTC m=+172.886914136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:34 crc kubenswrapper[4796]: I1127 11:27:34.969637 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:34 crc kubenswrapper[4796]: E1127 11:27:34.969862 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.469848028 +0000 UTC m=+172.988166946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.070648 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.070953 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.570942925 +0000 UTC m=+173.089261843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.160777 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.164381 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.164404 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.166038 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:35 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:35 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:35 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.166078 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.170416 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.171107 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.171228 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.671213691 +0000 UTC m=+173.189532609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.171408 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.171681 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.671673213 +0000 UTC m=+173.189992131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.235357 4796 generic.go:334] "Generic (PLEG): container finished" podID="71c9c343-afe0-4330-a97c-5600a14d569e" containerID="f1486a4a8ef526ab1bb55d3c90d9d65bc7f5e5c0d2ad18a2fa1d381bc91de2e1" exitCode=0 Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.235662 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71c9c343-afe0-4330-a97c-5600a14d569e","Type":"ContainerDied","Data":"f1486a4a8ef526ab1bb55d3c90d9d65bc7f5e5c0d2ad18a2fa1d381bc91de2e1"} Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.237368 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" event={"ID":"5f7f0a02-e94f-409c-8ed0-c5a7b301bace","Type":"ContainerStarted","Data":"2370d78918becca7c6e75389d87725b4926dd62181d78a908227a77b517a2e1b"} Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.240454 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.240506 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.240457 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.240736 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.251945 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhhp" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.272330 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.272479 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.772457721 +0000 UTC m=+173.290776639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.272770 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.273832 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.773820206 +0000 UTC m=+173.292139124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.292027 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.292370 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.293500 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.294010 4796 patch_prober.go:28] interesting pod/console-f9d7485db-bvccg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.294041 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bvccg" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.373395 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.373567 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.873534117 +0000 UTC m=+173.391853035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.373867 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.375111 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.875096247 +0000 UTC m=+173.393415165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.474824 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.475143 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:35.975127457 +0000 UTC m=+173.493446375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.503168 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cwptf" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.576443 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.578826 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.07881072 +0000 UTC m=+173.597129638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.664018 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.677623 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.677782 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.177751301 +0000 UTC m=+173.696070229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.678370 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.678739 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.178719447 +0000 UTC m=+173.697038435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.773151 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.779788 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.779992 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.279963417 +0000 UTC m=+173.798282335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.780299 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.780658 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.280650454 +0000 UTC m=+173.798969372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.880988 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.881402 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.381382972 +0000 UTC m=+173.899701890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.899121 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.899166 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.899997 4796 patch_prober.go:28] interesting pod/apiserver-76f77b778f-dp4q8 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.13:8443/livez\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.900040 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" podUID="95abf6de-142f-47bc-9492-38fc1e8f69b0" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.13:8443/livez\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.973668 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.981732 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:35 crc kubenswrapper[4796]: E1127 11:27:35.982700 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.482684304 +0000 UTC m=+174.001003222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.989583 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-zbnzg" Nov 27 11:27:35 crc kubenswrapper[4796]: I1127 11:27:35.993314 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.082384 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.083418 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.583402911 +0000 UTC m=+174.101721829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.092360 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g949p" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.140486 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6nhkv" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.143451 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.163244 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:36 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:36 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:36 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.163302 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.183700 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.185629 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.685618227 +0000 UTC m=+174.203937145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.242326 4796 generic.go:334] "Generic (PLEG): container finished" podID="9297600d-7a17-4a87-9ed1-26357ea3df51" containerID="299c7052f7d6546f9849981bf98b33114f042c5ea14a12fc07965930b9c95de6" exitCode=0 Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.242403 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" event={"ID":"9297600d-7a17-4a87-9ed1-26357ea3df51","Type":"ContainerDied","Data":"299c7052f7d6546f9849981bf98b33114f042c5ea14a12fc07965930b9c95de6"} Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.284640 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.285422 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.785397979 +0000 UTC m=+174.303716897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.385440 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.385802 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.885787148 +0000 UTC m=+174.404106066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.486068 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.486221 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.986198677 +0000 UTC m=+174.504517596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.486391 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.486654 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:36.986641789 +0000 UTC m=+174.504960707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.539290 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hxkmw"] Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.540165 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.543146 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.586882 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.587055 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr4ph\" (UniqueName: \"kubernetes.io/projected/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-kube-api-access-jr4ph\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.587099 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-catalog-content\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.587114 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-utilities\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.587229 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.087215952 +0000 UTC m=+174.605534870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.587539 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.600118 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxkmw"] Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.687930 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71c9c343-afe0-4330-a97c-5600a14d569e-kubelet-dir\") pod \"71c9c343-afe0-4330-a97c-5600a14d569e\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688019 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71c9c343-afe0-4330-a97c-5600a14d569e-kube-api-access\") pod \"71c9c343-afe0-4330-a97c-5600a14d569e\" (UID: \"71c9c343-afe0-4330-a97c-5600a14d569e\") " Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688308 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-utilities\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688351 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688413 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr4ph\" (UniqueName: \"kubernetes.io/projected/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-kube-api-access-jr4ph\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688455 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-catalog-content\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688816 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-catalog-content\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.688856 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71c9c343-afe0-4330-a97c-5600a14d569e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "71c9c343-afe0-4330-a97c-5600a14d569e" (UID: "71c9c343-afe0-4330-a97c-5600a14d569e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.690820 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.190796553 +0000 UTC m=+174.709115551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.690887 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-utilities\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.705924 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c9c343-afe0-4330-a97c-5600a14d569e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "71c9c343-afe0-4330-a97c-5600a14d569e" (UID: "71c9c343-afe0-4330-a97c-5600a14d569e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.716209 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr4ph\" (UniqueName: \"kubernetes.io/projected/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-kube-api-access-jr4ph\") pod \"community-operators-hxkmw\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.750070 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tddgt"] Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.750250 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c9c343-afe0-4330-a97c-5600a14d569e" containerName="pruner" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.750260 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c9c343-afe0-4330-a97c-5600a14d569e" containerName="pruner" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.750398 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c9c343-afe0-4330-a97c-5600a14d569e" containerName="pruner" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.752421 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.758115 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.780111 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tddgt"] Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.797172 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.797637 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbbks\" (UniqueName: \"kubernetes.io/projected/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-kube-api-access-nbbks\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.797723 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-catalog-content\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.797775 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-utilities\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.797815 4796 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71c9c343-afe0-4330-a97c-5600a14d569e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.797826 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71c9c343-afe0-4330-a97c-5600a14d569e-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.797888 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.297872863 +0000 UTC m=+174.816191781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.884525 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.901173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-catalog-content\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.901256 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.901301 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-utilities\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.901323 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbbks\" (UniqueName: \"kubernetes.io/projected/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-kube-api-access-nbbks\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.901954 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-catalog-content\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.902152 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-utilities\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: E1127 11:27:36.902257 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.402226313 +0000 UTC m=+174.920545231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.928507 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbbks\" (UniqueName: \"kubernetes.io/projected/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-kube-api-access-nbbks\") pod \"certified-operators-tddgt\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.930389 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-klnx9"] Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.931557 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:36 crc kubenswrapper[4796]: I1127 11:27:36.975320 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klnx9"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.002826 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.002905 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.502885828 +0000 UTC m=+175.021204736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.003348 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmvp7\" (UniqueName: \"kubernetes.io/projected/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-kube-api-access-qmvp7\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.003492 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-catalog-content\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.003813 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-utilities\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.003850 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.004181 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.504173842 +0000 UTC m=+175.022492770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.069020 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.104713 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.104832 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmvp7\" (UniqueName: \"kubernetes.io/projected/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-kube-api-access-qmvp7\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.104872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-catalog-content\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.104895 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-utilities\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.105317 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.605297839 +0000 UTC m=+175.123616747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.105630 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-utilities\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.105709 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-catalog-content\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.121533 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmvp7\" (UniqueName: \"kubernetes.io/projected/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-kube-api-access-qmvp7\") pod \"community-operators-klnx9\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.133466 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hrwn6"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.134596 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.141760 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrwn6"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.152522 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hxkmw"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.169437 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:37 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:37 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:37 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.169493 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.182550 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zs7tf" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.206234 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.206610 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.706597781 +0000 UTC m=+175.224916699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.245903 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.260852 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerStarted","Data":"a0b3651b70bbb09a86187f7e1bf649370fa0476c6063ee38259ee6d2e6b04c95"} Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.262960 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.263708 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71c9c343-afe0-4330-a97c-5600a14d569e","Type":"ContainerDied","Data":"98cfda4815d863acaa59c93e3618c212af9bfd8d888a5843e2489719b9592cbd"} Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.263743 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98cfda4815d863acaa59c93e3618c212af9bfd8d888a5843e2489719b9592cbd" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.307273 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.307434 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-utilities\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.307483 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-catalog-content\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.307533 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.807507973 +0000 UTC m=+175.325826891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.307575 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.307706 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b88q5\" (UniqueName: \"kubernetes.io/projected/244e65ef-52c0-49c2-b1dc-1c6e063d075b-kube-api-access-b88q5\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.308004 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.807997676 +0000 UTC m=+175.326316594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.405493 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tddgt"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.411237 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.411516 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-utilities\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.411559 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-catalog-content\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.411634 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.911598127 +0000 UTC m=+175.429917035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.411723 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.411832 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b88q5\" (UniqueName: \"kubernetes.io/projected/244e65ef-52c0-49c2-b1dc-1c6e063d075b-kube-api-access-b88q5\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.411951 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-catalog-content\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.412168 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-utilities\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.412454 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:37.912446409 +0000 UTC m=+175.430765327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.462140 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b88q5\" (UniqueName: \"kubernetes.io/projected/244e65ef-52c0-49c2-b1dc-1c6e063d075b-kube-api-access-b88q5\") pod \"certified-operators-hrwn6\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.512357 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.512752 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.012737534 +0000 UTC m=+175.531056452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.589404 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.590037 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.592241 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.595563 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klnx9"] Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.600863 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 11:27:37 crc kubenswrapper[4796]: W1127 11:27:37.608521 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddf010b4_d216_4ab7_bc54_a8879fcb45e0.slice/crio-bf761285ddd3db7ff3e3e0d329e91507b5ddd69eb398b57a503ea1332a6e741b WatchSource:0}: Error finding container bf761285ddd3db7ff3e3e0d329e91507b5ddd69eb398b57a503ea1332a6e741b: Status 404 returned error can't find the container with id bf761285ddd3db7ff3e3e0d329e91507b5ddd69eb398b57a503ea1332a6e741b Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.610536 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.614835 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.615135 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.115124305 +0000 UTC m=+175.633443223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.705694 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.715909 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.716328 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.716503 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.717058 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.217041952 +0000 UTC m=+175.735360870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.758680 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.818937 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9297600d-7a17-4a87-9ed1-26357ea3df51-config-volume\") pod \"9297600d-7a17-4a87-9ed1-26357ea3df51\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.819097 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9297600d-7a17-4a87-9ed1-26357ea3df51-secret-volume\") pod \"9297600d-7a17-4a87-9ed1-26357ea3df51\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.819120 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z64vn\" (UniqueName: \"kubernetes.io/projected/9297600d-7a17-4a87-9ed1-26357ea3df51-kube-api-access-z64vn\") pod \"9297600d-7a17-4a87-9ed1-26357ea3df51\" (UID: \"9297600d-7a17-4a87-9ed1-26357ea3df51\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.819232 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.819306 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.819327 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.819588 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.319574286 +0000 UTC m=+175.837893194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.820346 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9297600d-7a17-4a87-9ed1-26357ea3df51-config-volume" (OuterVolumeSpecName: "config-volume") pod "9297600d-7a17-4a87-9ed1-26357ea3df51" (UID: "9297600d-7a17-4a87-9ed1-26357ea3df51"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.826188 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.826634 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9297600d-7a17-4a87-9ed1-26357ea3df51-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9297600d-7a17-4a87-9ed1-26357ea3df51" (UID: "9297600d-7a17-4a87-9ed1-26357ea3df51"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.835238 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9297600d-7a17-4a87-9ed1-26357ea3df51-kube-api-access-z64vn" (OuterVolumeSpecName: "kube-api-access-z64vn") pod "9297600d-7a17-4a87-9ed1-26357ea3df51" (UID: "9297600d-7a17-4a87-9ed1-26357ea3df51"). InnerVolumeSpecName "kube-api-access-z64vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.842953 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.919614 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.919783 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9297600d-7a17-4a87-9ed1-26357ea3df51-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.919795 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9297600d-7a17-4a87-9ed1-26357ea3df51-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.919803 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z64vn\" (UniqueName: \"kubernetes.io/projected/9297600d-7a17-4a87-9ed1-26357ea3df51-kube-api-access-z64vn\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:37 crc kubenswrapper[4796]: E1127 11:27:37.919861 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.419848122 +0000 UTC m=+175.938167040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:37 crc kubenswrapper[4796]: I1127 11:27:37.920572 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.026456 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.026757 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.526745228 +0000 UTC m=+176.045064146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.127166 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.127370 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.627346132 +0000 UTC m=+176.145665050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.127548 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.128023 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.628007969 +0000 UTC m=+176.146326887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.166612 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:38 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:38 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:38 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.166686 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.194125 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 11:27:38 crc kubenswrapper[4796]: W1127 11:27:38.204172 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poddbef83c1_bd57_4488_9d37_f5c9b2d9e079.slice/crio-7ae5e27596a66bad1468fca6f5ce612b69de03e796eef6337cd4d9900cb09505 WatchSource:0}: Error finding container 7ae5e27596a66bad1468fca6f5ce612b69de03e796eef6337cd4d9900cb09505: Status 404 returned error can't find the container with id 7ae5e27596a66bad1468fca6f5ce612b69de03e796eef6337cd4d9900cb09505 Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.228226 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.228396 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.728370807 +0000 UTC m=+176.246689725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.228674 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.229005 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.728997603 +0000 UTC m=+176.247316521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.230713 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrwn6"] Nov 27 11:27:38 crc kubenswrapper[4796]: W1127 11:27:38.237155 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod244e65ef_52c0_49c2_b1dc_1c6e063d075b.slice/crio-118cb7834bb7b5fd79f924415132c49b510916dc7c63dda27568f45eb6a4fdbc WatchSource:0}: Error finding container 118cb7834bb7b5fd79f924415132c49b510916dc7c63dda27568f45eb6a4fdbc: Status 404 returned error can't find the container with id 118cb7834bb7b5fd79f924415132c49b510916dc7c63dda27568f45eb6a4fdbc Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.282299 4796 generic.go:334] "Generic (PLEG): container finished" podID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerID="a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86" exitCode=0 Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.282383 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerDied","Data":"a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.282408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerStarted","Data":"f0dcee177b78fc58050674ec6da03d213ccb7a9530f79e2b0e05fc1c60b5cafc"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.283989 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.286678 4796 generic.go:334] "Generic (PLEG): container finished" podID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerID="4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4" exitCode=0 Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.286759 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerDied","Data":"4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.292429 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" event={"ID":"9297600d-7a17-4a87-9ed1-26357ea3df51","Type":"ContainerDied","Data":"5567de5776d2865d2800413a5726681f9f30067285d4e145d19e64877c0d6f20"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.292473 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5567de5776d2865d2800413a5726681f9f30067285d4e145d19e64877c0d6f20" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.292508 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.293855 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerStarted","Data":"118cb7834bb7b5fd79f924415132c49b510916dc7c63dda27568f45eb6a4fdbc"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.308662 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dbef83c1-bd57-4488-9d37-f5c9b2d9e079","Type":"ContainerStarted","Data":"7ae5e27596a66bad1468fca6f5ce612b69de03e796eef6337cd4d9900cb09505"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.314757 4796 generic.go:334] "Generic (PLEG): container finished" podID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerID="13ff7d0016887d4175a933baa62bc6d90d6469f1ddf19863671dce6d8c9eea7a" exitCode=0 Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.314803 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerDied","Data":"13ff7d0016887d4175a933baa62bc6d90d6469f1ddf19863671dce6d8c9eea7a"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.315023 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerStarted","Data":"bf761285ddd3db7ff3e3e0d329e91507b5ddd69eb398b57a503ea1332a6e741b"} Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.329225 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.329625 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.829609547 +0000 UTC m=+176.347928465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.431327 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.431721 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:38.931704179 +0000 UTC m=+176.450023097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.532402 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.532812 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.032777416 +0000 UTC m=+176.551096364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.634235 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.634659 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.134637352 +0000 UTC m=+176.652956280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.729470 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pbl4f"] Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.729734 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9297600d-7a17-4a87-9ed1-26357ea3df51" containerName="collect-profiles" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.729754 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9297600d-7a17-4a87-9ed1-26357ea3df51" containerName="collect-profiles" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.729898 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9297600d-7a17-4a87-9ed1-26357ea3df51" containerName="collect-profiles" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.730776 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.735431 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.735568 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.235548684 +0000 UTC m=+176.753867602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.735780 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-utilities\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.735860 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqrfj\" (UniqueName: \"kubernetes.io/projected/08792b59-9118-40df-ab6a-ef002d223f4e-kube-api-access-vqrfj\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.735967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.736105 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.736235 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.236227571 +0000 UTC m=+176.754546489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.736285 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-catalog-content\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.744506 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbl4f"] Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.836999 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.837169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-utilities\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.837208 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.337177294 +0000 UTC m=+176.855496222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.837340 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqrfj\" (UniqueName: \"kubernetes.io/projected/08792b59-9118-40df-ab6a-ef002d223f4e-kube-api-access-vqrfj\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.837455 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.837498 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-catalog-content\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.837721 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-utilities\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.838050 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-catalog-content\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.838070 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.338052297 +0000 UTC m=+176.856371225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.880566 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqrfj\" (UniqueName: \"kubernetes.io/projected/08792b59-9118-40df-ab6a-ef002d223f4e-kube-api-access-vqrfj\") pod \"redhat-marketplace-pbl4f\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.937983 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.938113 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.438086646 +0000 UTC m=+176.956405574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:38 crc kubenswrapper[4796]: I1127 11:27:38.938165 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:38 crc kubenswrapper[4796]: E1127 11:27:38.938444 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.438432905 +0000 UTC m=+176.956751823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.038912 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.039079 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.53906194 +0000 UTC m=+177.057380858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.039127 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.039465 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.53945337 +0000 UTC m=+177.057772288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.045496 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.141549 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.141992 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.641979293 +0000 UTC m=+177.160298211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.146359 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-djbsd"] Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.147707 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.153788 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-djbsd"] Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.171305 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:39 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:39 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:39 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.171354 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.244210 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.244861 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.744848486 +0000 UTC m=+177.263167404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.261165 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbl4f"] Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.320795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerStarted","Data":"365ee2422dd27d8a89ff2ac83a738c141278852cd4bddfd5d7896ee0a282339e"} Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.321996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dbef83c1-bd57-4488-9d37-f5c9b2d9e079","Type":"ContainerStarted","Data":"649a48bd53987805f7f9349029ad9dbc5f370f2a2fca987ce0c3cc3da62e283d"} Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.325045 4796 generic.go:334] "Generic (PLEG): container finished" podID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerID="42c764fc7848cd120b2f5dfd696dcade5e79e264e543f28cf75c02ac23ffc25c" exitCode=0 Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.325094 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerDied","Data":"42c764fc7848cd120b2f5dfd696dcade5e79e264e543f28cf75c02ac23ffc25c"} Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.327780 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" event={"ID":"5f7f0a02-e94f-409c-8ed0-c5a7b301bace","Type":"ContainerStarted","Data":"d11f96218b21fcc1e1cec7846d86f7925c84202c38239413acbf58f83f6a5ea2"} Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.335278 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.335252668 podStartE2EDuration="2.335252668s" podCreationTimestamp="2025-11-27 11:27:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:39.33380146 +0000 UTC m=+176.852120378" watchObservedRunningTime="2025-11-27 11:27:39.335252668 +0000 UTC m=+176.853571586" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.345587 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.345724 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.845708397 +0000 UTC m=+177.364027315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.345870 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.345899 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljsm2\" (UniqueName: \"kubernetes.io/projected/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-kube-api-access-ljsm2\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.345927 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-catalog-content\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.346238 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.84622882 +0000 UTC m=+177.364547738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.346393 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-utilities\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.447769 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.448013 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.947980483 +0000 UTC m=+177.466299411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.448233 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-utilities\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.448407 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.448449 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljsm2\" (UniqueName: \"kubernetes.io/projected/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-kube-api-access-ljsm2\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.448497 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-catalog-content\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.448825 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:39.948806875 +0000 UTC m=+177.467125803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.448912 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-utilities\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.448970 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-catalog-content\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.475897 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljsm2\" (UniqueName: \"kubernetes.io/projected/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-kube-api-access-ljsm2\") pod \"redhat-marketplace-djbsd\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.488488 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.549437 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.549612 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:40.049583444 +0000 UTC m=+177.567902452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.549774 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.550075 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:40.050061376 +0000 UTC m=+177.568380304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.571935 4796 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.650545 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.651007 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 11:27:40.150992338 +0000 UTC m=+177.669311256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.730254 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-92gc9"] Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.731766 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.735230 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.737007 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-92gc9"] Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.752164 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-utilities\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.752231 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksnfm\" (UniqueName: \"kubernetes.io/projected/906ed244-7663-473c-8495-fe03bd4aa745-kube-api-access-ksnfm\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.752342 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-catalog-content\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.752390 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: E1127 11:27:39.752674 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 11:27:40.252662459 +0000 UTC m=+177.770981387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5brmj" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.766789 4796 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-27T11:27:39.571959228Z","Handler":null,"Name":""} Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.770819 4796 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.770858 4796 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.853043 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.853397 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksnfm\" (UniqueName: \"kubernetes.io/projected/906ed244-7663-473c-8495-fe03bd4aa745-kube-api-access-ksnfm\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.853475 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-catalog-content\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.853635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-utilities\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.853959 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-catalog-content\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.854207 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-utilities\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.858168 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.871386 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksnfm\" (UniqueName: \"kubernetes.io/projected/906ed244-7663-473c-8495-fe03bd4aa745-kube-api-access-ksnfm\") pod \"redhat-operators-92gc9\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.926226 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-djbsd"] Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.954366 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.957070 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 11:27:39 crc kubenswrapper[4796]: I1127 11:27:39.957365 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.054238 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:27:40 crc kubenswrapper[4796]: W1127 11:27:40.120876 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59533ee5_3cf6_43d4_a75e_9611c9c84cb2.slice/crio-39f47c0a521e0a2b51a2f25a3e36b1a204c44f29476a543dbd0358d4b024aef2 WatchSource:0}: Error finding container 39f47c0a521e0a2b51a2f25a3e36b1a204c44f29476a543dbd0358d4b024aef2: Status 404 returned error can't find the container with id 39f47c0a521e0a2b51a2f25a3e36b1a204c44f29476a543dbd0358d4b024aef2 Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.139494 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2j2sz"] Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.142241 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.150209 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2j2sz"] Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.158154 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-utilities\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.158256 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-catalog-content\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.158308 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsbmn\" (UniqueName: \"kubernetes.io/projected/6ee15471-b388-4e39-adca-a7409e0e90be-kube-api-access-gsbmn\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.182429 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:40 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:40 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:40 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.182502 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.246409 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5brmj\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.258768 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-utilities\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.258873 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-catalog-content\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.258909 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsbmn\" (UniqueName: \"kubernetes.io/projected/6ee15471-b388-4e39-adca-a7409e0e90be-kube-api-access-gsbmn\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.259533 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-utilities\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.259749 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-catalog-content\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.278542 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsbmn\" (UniqueName: \"kubernetes.io/projected/6ee15471-b388-4e39-adca-a7409e0e90be-kube-api-access-gsbmn\") pod \"redhat-operators-2j2sz\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.301753 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.338317 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" event={"ID":"5f7f0a02-e94f-409c-8ed0-c5a7b301bace","Type":"ContainerStarted","Data":"a70f8cd668d66cfb9fc6bb8e130c9f78920e98e37898980573c4d3ccd7b75628"} Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.339752 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerStarted","Data":"39f47c0a521e0a2b51a2f25a3e36b1a204c44f29476a543dbd0358d4b024aef2"} Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.394440 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-92gc9"] Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.405788 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.724222 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2j2sz"] Nov 27 11:27:40 crc kubenswrapper[4796]: W1127 11:27:40.728716 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ee15471_b388_4e39_adca_a7409e0e90be.slice/crio-e512b060ec2304d67578336b3b024e4ee12c873751b513649931da5beccf05c5 WatchSource:0}: Error finding container e512b060ec2304d67578336b3b024e4ee12c873751b513649931da5beccf05c5: Status 404 returned error can't find the container with id e512b060ec2304d67578336b3b024e4ee12c873751b513649931da5beccf05c5 Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.864891 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5brmj"] Nov 27 11:27:40 crc kubenswrapper[4796]: W1127 11:27:40.871819 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e1b281d_0c21_49d8_89fd_ec2534405cdf.slice/crio-1f774e9f4c37dc25be8d05d91083f70b3e3d8d70d46ac376d7b5aaa666aedba1 WatchSource:0}: Error finding container 1f774e9f4c37dc25be8d05d91083f70b3e3d8d70d46ac376d7b5aaa666aedba1: Status 404 returned error can't find the container with id 1f774e9f4c37dc25be8d05d91083f70b3e3d8d70d46ac376d7b5aaa666aedba1 Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.904329 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:40 crc kubenswrapper[4796]: I1127 11:27:40.911944 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-dp4q8" Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.167535 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:41 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:41 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:41 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.167585 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.196234 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2vdl8" Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.348854 4796 generic.go:334] "Generic (PLEG): container finished" podID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerID="1ebab4aaf456e5d5b3195cf733547d4ad92691175f6c838daad48b884a7d4c8e" exitCode=0 Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.348950 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerDied","Data":"1ebab4aaf456e5d5b3195cf733547d4ad92691175f6c838daad48b884a7d4c8e"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.350393 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbef83c1-bd57-4488-9d37-f5c9b2d9e079" containerID="649a48bd53987805f7f9349029ad9dbc5f370f2a2fca987ce0c3cc3da62e283d" exitCode=0 Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.350455 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dbef83c1-bd57-4488-9d37-f5c9b2d9e079","Type":"ContainerDied","Data":"649a48bd53987805f7f9349029ad9dbc5f370f2a2fca987ce0c3cc3da62e283d"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.358698 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" event={"ID":"5f7f0a02-e94f-409c-8ed0-c5a7b301bace","Type":"ContainerStarted","Data":"a757310dfd88ac6f2be0064368bc3c59049562fa42f7f72e632b1c7a6f37b597"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.360206 4796 generic.go:334] "Generic (PLEG): container finished" podID="08792b59-9118-40df-ab6a-ef002d223f4e" containerID="b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4" exitCode=0 Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.360251 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerDied","Data":"b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.363730 4796 generic.go:334] "Generic (PLEG): container finished" podID="906ed244-7663-473c-8495-fe03bd4aa745" containerID="3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d" exitCode=0 Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.363774 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerDied","Data":"3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.363791 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerStarted","Data":"3fda918261abca918ce66ce611448f7f1aaeadc87f8ae13249930497eda9fdc7"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.369866 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" event={"ID":"5e1b281d-0c21-49d8-89fd-ec2534405cdf","Type":"ContainerStarted","Data":"158891edaf4dfc616e75d3cd2bd662a321da935c0880abcc34f57648af671a4a"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.369918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" event={"ID":"5e1b281d-0c21-49d8-89fd-ec2534405cdf","Type":"ContainerStarted","Data":"1f774e9f4c37dc25be8d05d91083f70b3e3d8d70d46ac376d7b5aaa666aedba1"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.372511 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ee15471-b388-4e39-adca-a7409e0e90be" containerID="9c35dd07466aadf358ceedefd562d8a4374afe88c3b3a3a4f706257a30594b22" exitCode=0 Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.373528 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerDied","Data":"9c35dd07466aadf358ceedefd562d8a4374afe88c3b3a3a4f706257a30594b22"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.373552 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerStarted","Data":"e512b060ec2304d67578336b3b024e4ee12c873751b513649931da5beccf05c5"} Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.469219 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7t7jg" podStartSLOduration=19.46919791 podStartE2EDuration="19.46919791s" podCreationTimestamp="2025-11-27 11:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:41.468073151 +0000 UTC m=+178.986392069" watchObservedRunningTime="2025-11-27 11:27:41.46919791 +0000 UTC m=+178.987516828" Nov 27 11:27:41 crc kubenswrapper[4796]: I1127 11:27:41.577846 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.173966 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:42 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:42 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:42 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.174033 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.377776 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.605843 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.618652 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" podStartSLOduration=158.618636594 podStartE2EDuration="2m38.618636594s" podCreationTimestamp="2025-11-27 11:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:27:42.395320138 +0000 UTC m=+179.913639076" watchObservedRunningTime="2025-11-27 11:27:42.618636594 +0000 UTC m=+180.136955502" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.621258 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kube-api-access\") pod \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.621384 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kubelet-dir\") pod \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\" (UID: \"dbef83c1-bd57-4488-9d37-f5c9b2d9e079\") " Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.640124 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dbef83c1-bd57-4488-9d37-f5c9b2d9e079" (UID: "dbef83c1-bd57-4488-9d37-f5c9b2d9e079"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.640665 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dbef83c1-bd57-4488-9d37-f5c9b2d9e079" (UID: "dbef83c1-bd57-4488-9d37-f5c9b2d9e079"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.723389 4796 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:42 crc kubenswrapper[4796]: I1127 11:27:42.723419 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbef83c1-bd57-4488-9d37-f5c9b2d9e079-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:27:43 crc kubenswrapper[4796]: I1127 11:27:43.163061 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:43 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:43 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:43 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:43 crc kubenswrapper[4796]: I1127 11:27:43.163129 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:43 crc kubenswrapper[4796]: I1127 11:27:43.384898 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"dbef83c1-bd57-4488-9d37-f5c9b2d9e079","Type":"ContainerDied","Data":"7ae5e27596a66bad1468fca6f5ce612b69de03e796eef6337cd4d9900cb09505"} Nov 27 11:27:43 crc kubenswrapper[4796]: I1127 11:27:43.384955 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ae5e27596a66bad1468fca6f5ce612b69de03e796eef6337cd4d9900cb09505" Nov 27 11:27:43 crc kubenswrapper[4796]: I1127 11:27:43.385215 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 11:27:44 crc kubenswrapper[4796]: I1127 11:27:44.163033 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:44 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:44 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:44 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:44 crc kubenswrapper[4796]: I1127 11:27:44.163513 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.162883 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:45 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:45 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:45 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.163252 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.240415 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.240472 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.240483 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.240582 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.292006 4796 patch_prober.go:28] interesting pod/console-f9d7485db-bvccg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 27 11:27:45 crc kubenswrapper[4796]: I1127 11:27:45.292055 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bvccg" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 27 11:27:46 crc kubenswrapper[4796]: I1127 11:27:46.163220 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:46 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:46 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:46 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:46 crc kubenswrapper[4796]: I1127 11:27:46.163316 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:47 crc kubenswrapper[4796]: I1127 11:27:47.162749 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:47 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:47 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:47 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:47 crc kubenswrapper[4796]: I1127 11:27:47.162804 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:48 crc kubenswrapper[4796]: I1127 11:27:48.163549 4796 patch_prober.go:28] interesting pod/router-default-5444994796-v65j9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 11:27:48 crc kubenswrapper[4796]: [-]has-synced failed: reason withheld Nov 27 11:27:48 crc kubenswrapper[4796]: [+]process-running ok Nov 27 11:27:48 crc kubenswrapper[4796]: healthz check failed Nov 27 11:27:48 crc kubenswrapper[4796]: I1127 11:27:48.163631 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v65j9" podUID="b9b0d1de-c306-49d7-a94e-c83c054e53ef" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 11:27:49 crc kubenswrapper[4796]: I1127 11:27:49.163091 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:49 crc kubenswrapper[4796]: I1127 11:27:49.164963 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-v65j9" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.249543 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.249616 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.250161 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.250225 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.250299 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.250882 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"9ed7cf3c85d2db4f966ce68e80da16ffb2929901542f984eff7cd9a76d2e569e"} pod="openshift-console/downloads-7954f5f757-rdzj8" containerMessage="Container download-server failed liveness probe, will be restarted" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.250980 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" containerID="cri-o://9ed7cf3c85d2db4f966ce68e80da16ffb2929901542f984eff7cd9a76d2e569e" gracePeriod=2 Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.252700 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.252747 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.304286 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:55 crc kubenswrapper[4796]: I1127 11:27:55.311876 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:27:56 crc kubenswrapper[4796]: I1127 11:27:56.463790 4796 generic.go:334] "Generic (PLEG): container finished" podID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerID="9ed7cf3c85d2db4f966ce68e80da16ffb2929901542f984eff7cd9a76d2e569e" exitCode=0 Nov 27 11:27:56 crc kubenswrapper[4796]: I1127 11:27:56.463918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rdzj8" event={"ID":"ce3173af-2480-4a96-a7c5-ab9c44b18c64","Type":"ContainerDied","Data":"9ed7cf3c85d2db4f966ce68e80da16ffb2929901542f984eff7cd9a76d2e569e"} Nov 27 11:28:00 crc kubenswrapper[4796]: I1127 11:28:00.412901 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:28:01 crc kubenswrapper[4796]: I1127 11:28:01.879592 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:28:01 crc kubenswrapper[4796]: I1127 11:28:01.879706 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:28:05 crc kubenswrapper[4796]: I1127 11:28:05.250585 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:28:05 crc kubenswrapper[4796]: I1127 11:28:05.251010 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:28:06 crc kubenswrapper[4796]: I1127 11:28:06.149821 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pzxv6" Nov 27 11:28:15 crc kubenswrapper[4796]: I1127 11:28:15.241696 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:28:15 crc kubenswrapper[4796]: I1127 11:28:15.242181 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.976128 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 11:28:18 crc kubenswrapper[4796]: E1127 11:28:18.976921 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbef83c1-bd57-4488-9d37-f5c9b2d9e079" containerName="pruner" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.976944 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbef83c1-bd57-4488-9d37-f5c9b2d9e079" containerName="pruner" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.977111 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbef83c1-bd57-4488-9d37-f5c9b2d9e079" containerName="pruner" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.977823 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.983981 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.984630 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 11:28:18 crc kubenswrapper[4796]: I1127 11:28:18.991975 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.044593 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5ce83c0-1f80-485d-970b-51beaab3f6df-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.044684 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5ce83c0-1f80-485d-970b-51beaab3f6df-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.145399 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5ce83c0-1f80-485d-970b-51beaab3f6df-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.145512 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5ce83c0-1f80-485d-970b-51beaab3f6df-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.145832 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5ce83c0-1f80-485d-970b-51beaab3f6df-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.182257 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5ce83c0-1f80-485d-970b-51beaab3f6df-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:19 crc kubenswrapper[4796]: I1127 11:28:19.314785 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:28:23 crc kubenswrapper[4796]: I1127 11:28:23.981605 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 11:28:23 crc kubenswrapper[4796]: I1127 11:28:23.983448 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:23 crc kubenswrapper[4796]: I1127 11:28:23.999541 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.123718 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99b0a522-adc4-4078-ad97-941f1597a69c-kube-api-access\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.123871 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-var-lock\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.123973 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.225156 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99b0a522-adc4-4078-ad97-941f1597a69c-kube-api-access\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.225347 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-var-lock\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.225522 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.225629 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-var-lock\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.225685 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.871607 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99b0a522-adc4-4078-ad97-941f1597a69c-kube-api-access\") pod \"installer-9-crc\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:24 crc kubenswrapper[4796]: I1127 11:28:24.928832 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:28:25 crc kubenswrapper[4796]: I1127 11:28:25.240847 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:28:25 crc kubenswrapper[4796]: I1127 11:28:25.240913 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:28:31 crc kubenswrapper[4796]: I1127 11:28:31.879050 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:28:31 crc kubenswrapper[4796]: I1127 11:28:31.879904 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:28:31 crc kubenswrapper[4796]: I1127 11:28:31.880005 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:28:31 crc kubenswrapper[4796]: I1127 11:28:31.881090 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:28:31 crc kubenswrapper[4796]: I1127 11:28:31.881198 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e" gracePeriod=600 Nov 27 11:28:35 crc kubenswrapper[4796]: I1127 11:28:35.241065 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:28:35 crc kubenswrapper[4796]: I1127 11:28:35.241731 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:28:39 crc kubenswrapper[4796]: I1127 11:28:39.726625 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e" exitCode=0 Nov 27 11:28:39 crc kubenswrapper[4796]: I1127 11:28:39.726715 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e"} Nov 27 11:28:41 crc kubenswrapper[4796]: E1127 11:28:41.329137 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 27 11:28:41 crc kubenswrapper[4796]: E1127 11:28:41.329334 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nbbks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tddgt_openshift-marketplace(a2bcf5d3-efae-40cb-b701-fe0f65a28d0d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:28:41 crc kubenswrapper[4796]: E1127 11:28:41.331097 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tddgt" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" Nov 27 11:28:45 crc kubenswrapper[4796]: I1127 11:28:45.241607 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:28:45 crc kubenswrapper[4796]: I1127 11:28:45.241980 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:28:46 crc kubenswrapper[4796]: E1127 11:28:46.508333 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tddgt" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" Nov 27 11:28:47 crc kubenswrapper[4796]: E1127 11:28:47.143121 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 27 11:28:47 crc kubenswrapper[4796]: E1127 11:28:47.143643 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmvp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-klnx9_openshift-marketplace(ddf010b4-d216-4ab7-bc54-a8879fcb45e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:28:47 crc kubenswrapper[4796]: E1127 11:28:47.144904 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-klnx9" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" Nov 27 11:28:55 crc kubenswrapper[4796]: I1127 11:28:55.240889 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:28:55 crc kubenswrapper[4796]: I1127 11:28:55.241597 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:28:55 crc kubenswrapper[4796]: E1127 11:28:55.846619 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 27 11:28:55 crc kubenswrapper[4796]: E1127 11:28:55.846763 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jr4ph,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hxkmw_openshift-marketplace(83a6d9b2-0bc7-46fb-9528-afce9582aaaf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:28:55 crc kubenswrapper[4796]: E1127 11:28:55.847969 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hxkmw" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" Nov 27 11:29:00 crc kubenswrapper[4796]: E1127 11:29:00.036631 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hxkmw" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" Nov 27 11:29:00 crc kubenswrapper[4796]: E1127 11:29:00.071785 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 27 11:29:00 crc kubenswrapper[4796]: E1127 11:29:00.071956 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gsbmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2j2sz_openshift-marketplace(6ee15471-b388-4e39-adca-a7409e0e90be): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:29:00 crc kubenswrapper[4796]: E1127 11:29:00.073186 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2j2sz" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" Nov 27 11:29:00 crc kubenswrapper[4796]: I1127 11:29:00.526872 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 11:29:00 crc kubenswrapper[4796]: I1127 11:29:00.579818 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 11:29:00 crc kubenswrapper[4796]: W1127 11:29:00.598335 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod99b0a522_adc4_4078_ad97_941f1597a69c.slice/crio-cacadfa6ac3d70549e14d7249bc08982fddb61c79832060db2939c930cdc9c9b WatchSource:0}: Error finding container cacadfa6ac3d70549e14d7249bc08982fddb61c79832060db2939c930cdc9c9b: Status 404 returned error can't find the container with id cacadfa6ac3d70549e14d7249bc08982fddb61c79832060db2939c930cdc9c9b Nov 27 11:29:00 crc kubenswrapper[4796]: I1127 11:29:00.862994 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b5ce83c0-1f80-485d-970b-51beaab3f6df","Type":"ContainerStarted","Data":"fdd3278a3725a3bd9f5b2107aaa90ceef029a2f5c7700ed7233f0d46ac1fc71d"} Nov 27 11:29:00 crc kubenswrapper[4796]: I1127 11:29:00.864487 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"99b0a522-adc4-4078-ad97-941f1597a69c","Type":"ContainerStarted","Data":"cacadfa6ac3d70549e14d7249bc08982fddb61c79832060db2939c930cdc9c9b"} Nov 27 11:29:00 crc kubenswrapper[4796]: E1127 11:29:00.866686 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2j2sz" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" Nov 27 11:29:03 crc kubenswrapper[4796]: I1127 11:29:03.893162 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b5ce83c0-1f80-485d-970b-51beaab3f6df","Type":"ContainerStarted","Data":"47d1386079b925e053a0b776b518c8c968506a0028bf6311f485966692ff7e53"} Nov 27 11:29:03 crc kubenswrapper[4796]: I1127 11:29:03.895897 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rdzj8" event={"ID":"ce3173af-2480-4a96-a7c5-ab9c44b18c64","Type":"ContainerStarted","Data":"6a2271859ab3b2ccf111f8174a8b52b16113ef0e354960ce74508a8b78a76e6c"} Nov 27 11:29:04 crc kubenswrapper[4796]: I1127 11:29:04.906543 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"2bc2f5728589faed16620248b20e0a440b5a67ed778b4753d5c8d2dca98cb558"} Nov 27 11:29:04 crc kubenswrapper[4796]: I1127 11:29:04.911658 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"99b0a522-adc4-4078-ad97-941f1597a69c","Type":"ContainerStarted","Data":"4b6cc76710e99ed37095732fc5cefa9cbcdb99521399bf3943abd959301e845d"} Nov 27 11:29:04 crc kubenswrapper[4796]: I1127 11:29:04.912343 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:29:04 crc kubenswrapper[4796]: I1127 11:29:04.912579 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:29:04 crc kubenswrapper[4796]: I1127 11:29:04.912626 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:29:05 crc kubenswrapper[4796]: E1127 11:29:05.042846 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 27 11:29:05 crc kubenswrapper[4796]: E1127 11:29:05.043339 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ksnfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-92gc9_openshift-marketplace(906ed244-7663-473c-8495-fe03bd4aa745): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:29:05 crc kubenswrapper[4796]: E1127 11:29:05.044714 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-92gc9" podUID="906ed244-7663-473c-8495-fe03bd4aa745" Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.241518 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.242877 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.241529 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.243783 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.917323 4796 generic.go:334] "Generic (PLEG): container finished" podID="b5ce83c0-1f80-485d-970b-51beaab3f6df" containerID="47d1386079b925e053a0b776b518c8c968506a0028bf6311f485966692ff7e53" exitCode=0 Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.917425 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b5ce83c0-1f80-485d-970b-51beaab3f6df","Type":"ContainerDied","Data":"47d1386079b925e053a0b776b518c8c968506a0028bf6311f485966692ff7e53"} Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.917925 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.917968 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:29:05 crc kubenswrapper[4796]: I1127 11:29:05.946199 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=42.946170569 podStartE2EDuration="42.946170569s" podCreationTimestamp="2025-11-27 11:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:29:04.959873951 +0000 UTC m=+262.478192909" watchObservedRunningTime="2025-11-27 11:29:05.946170569 +0000 UTC m=+263.464489497" Nov 27 11:29:08 crc kubenswrapper[4796]: E1127 11:29:08.563209 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-92gc9" podUID="906ed244-7663-473c-8495-fe03bd4aa745" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.631105 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.766695 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5ce83c0-1f80-485d-970b-51beaab3f6df-kubelet-dir\") pod \"b5ce83c0-1f80-485d-970b-51beaab3f6df\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.766847 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5ce83c0-1f80-485d-970b-51beaab3f6df-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b5ce83c0-1f80-485d-970b-51beaab3f6df" (UID: "b5ce83c0-1f80-485d-970b-51beaab3f6df"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.766930 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5ce83c0-1f80-485d-970b-51beaab3f6df-kube-api-access\") pod \"b5ce83c0-1f80-485d-970b-51beaab3f6df\" (UID: \"b5ce83c0-1f80-485d-970b-51beaab3f6df\") " Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.767300 4796 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5ce83c0-1f80-485d-970b-51beaab3f6df-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.773230 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5ce83c0-1f80-485d-970b-51beaab3f6df-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b5ce83c0-1f80-485d-970b-51beaab3f6df" (UID: "b5ce83c0-1f80-485d-970b-51beaab3f6df"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.868442 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5ce83c0-1f80-485d-970b-51beaab3f6df-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.944103 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b5ce83c0-1f80-485d-970b-51beaab3f6df","Type":"ContainerDied","Data":"fdd3278a3725a3bd9f5b2107aaa90ceef029a2f5c7700ed7233f0d46ac1fc71d"} Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.944164 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd3278a3725a3bd9f5b2107aaa90ceef029a2f5c7700ed7233f0d46ac1fc71d" Nov 27 11:29:08 crc kubenswrapper[4796]: I1127 11:29:08.944211 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 11:29:12 crc kubenswrapper[4796]: E1127 11:29:12.485421 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 11:29:12 crc kubenswrapper[4796]: E1127 11:29:12.485577 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vqrfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pbl4f_openshift-marketplace(08792b59-9118-40df-ab6a-ef002d223f4e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:29:12 crc kubenswrapper[4796]: E1127 11:29:12.486733 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pbl4f" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" Nov 27 11:29:13 crc kubenswrapper[4796]: E1127 11:29:13.619322 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 27 11:29:13 crc kubenswrapper[4796]: E1127 11:29:13.619503 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b88q5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hrwn6_openshift-marketplace(244e65ef-52c0-49c2-b1dc-1c6e063d075b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:29:13 crc kubenswrapper[4796]: E1127 11:29:13.620833 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hrwn6" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.646762 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.646808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.646864 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.649573 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.650559 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.650566 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.660128 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.664578 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.673772 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.748191 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.752940 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.897253 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.906435 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.973631 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:29:13 crc kubenswrapper[4796]: I1127 11:29:13.989406 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 11:29:15 crc kubenswrapper[4796]: I1127 11:29:15.240669 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:29:15 crc kubenswrapper[4796]: I1127 11:29:15.240813 4796 patch_prober.go:28] interesting pod/downloads-7954f5f757-rdzj8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 27 11:29:15 crc kubenswrapper[4796]: I1127 11:29:15.240888 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:29:15 crc kubenswrapper[4796]: I1127 11:29:15.240812 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rdzj8" podUID="ce3173af-2480-4a96-a7c5-ab9c44b18c64" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 27 11:29:21 crc kubenswrapper[4796]: E1127 11:29:19.490321 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 11:29:21 crc kubenswrapper[4796]: E1127 11:29:19.490952 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljsm2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-djbsd_openshift-marketplace(59533ee5-3cf6-43d4-a75e-9611c9c84cb2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 11:29:21 crc kubenswrapper[4796]: E1127 11:29:19.492337 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-djbsd" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" Nov 27 11:29:23 crc kubenswrapper[4796]: E1127 11:29:23.635677 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-djbsd" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" Nov 27 11:29:24 crc kubenswrapper[4796]: I1127 11:29:24.061748 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerStarted","Data":"bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0"} Nov 27 11:29:24 crc kubenswrapper[4796]: I1127 11:29:24.063530 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerStarted","Data":"4a81f02b197ad0b7c58a8e561ddba5107885352bb1a5a1b8b6b877e8cd291ec7"} Nov 27 11:29:24 crc kubenswrapper[4796]: W1127 11:29:24.130497 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-0413feb4ee191170f98d9cbfdb0c2c7307d57f12b20662ef0824e6ebe94e4c59 WatchSource:0}: Error finding container 0413feb4ee191170f98d9cbfdb0c2c7307d57f12b20662ef0824e6ebe94e4c59: Status 404 returned error can't find the container with id 0413feb4ee191170f98d9cbfdb0c2c7307d57f12b20662ef0824e6ebe94e4c59 Nov 27 11:29:24 crc kubenswrapper[4796]: W1127 11:29:24.218553 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-a02549999512fe213dc46ab86dee516e48fd0c227d0b0004ee76c65b8e64a210 WatchSource:0}: Error finding container a02549999512fe213dc46ab86dee516e48fd0c227d0b0004ee76c65b8e64a210: Status 404 returned error can't find the container with id a02549999512fe213dc46ab86dee516e48fd0c227d0b0004ee76c65b8e64a210 Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.072795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fb5e78668f2a86fd628cf0072f5a90a5137031ff4925fefe8eca40938ca4e7d2"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.072868 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a02549999512fe213dc46ab86dee516e48fd0c227d0b0004ee76c65b8e64a210"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.075480 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d59b711f9975750ad42e4777a6bbe967301100f584ec30799942051f4b10f53d"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.075539 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2bd8e301cb40fc63e3fd13bf3b66340b41cd822edca679a1f1ba157d7d56ebbb"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.075764 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.078753 4796 generic.go:334] "Generic (PLEG): container finished" podID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerID="4a81f02b197ad0b7c58a8e561ddba5107885352bb1a5a1b8b6b877e8cd291ec7" exitCode=0 Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.079005 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerDied","Data":"4a81f02b197ad0b7c58a8e561ddba5107885352bb1a5a1b8b6b877e8cd291ec7"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.082125 4796 generic.go:334] "Generic (PLEG): container finished" podID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerID="bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0" exitCode=0 Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.082228 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerDied","Data":"bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.084860 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c41ffe2c744d7db5a0011e4ce0a367c1b0e107cd21ea7854bdc596693ac276a3"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.084906 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0413feb4ee191170f98d9cbfdb0c2c7307d57f12b20662ef0824e6ebe94e4c59"} Nov 27 11:29:25 crc kubenswrapper[4796]: I1127 11:29:25.256538 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rdzj8" Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.108817 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerStarted","Data":"a1b81c6a6daa4769044a0ddaa6ed3b628e346b4e7d29adc1d5ba32870b43e902"} Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.110791 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerStarted","Data":"dde1bf859ced632eff9237f78321337b0adea5db7e88917249c1a008daab9e7e"} Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.112340 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerStarted","Data":"b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa"} Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.115510 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerStarted","Data":"eb49f5ec329482b9e62a2b40279bc18f02700a2762f9524eb9c0f1cc8e53478b"} Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.117772 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerStarted","Data":"8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6"} Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.119710 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerStarted","Data":"65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020"} Nov 27 11:29:28 crc kubenswrapper[4796]: I1127 11:29:28.122032 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerStarted","Data":"4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab"} Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.129521 4796 generic.go:334] "Generic (PLEG): container finished" podID="906ed244-7663-473c-8495-fe03bd4aa745" containerID="65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020" exitCode=0 Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.129623 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerDied","Data":"65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020"} Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.134086 4796 generic.go:334] "Generic (PLEG): container finished" podID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerID="4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab" exitCode=0 Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.134158 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerDied","Data":"4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab"} Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.138878 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ee15471-b388-4e39-adca-a7409e0e90be" containerID="a1b81c6a6daa4769044a0ddaa6ed3b628e346b4e7d29adc1d5ba32870b43e902" exitCode=0 Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.139164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerDied","Data":"a1b81c6a6daa4769044a0ddaa6ed3b628e346b4e7d29adc1d5ba32870b43e902"} Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.141398 4796 generic.go:334] "Generic (PLEG): container finished" podID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerID="dde1bf859ced632eff9237f78321337b0adea5db7e88917249c1a008daab9e7e" exitCode=0 Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.141537 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerDied","Data":"dde1bf859ced632eff9237f78321337b0adea5db7e88917249c1a008daab9e7e"} Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.145010 4796 generic.go:334] "Generic (PLEG): container finished" podID="08792b59-9118-40df-ab6a-ef002d223f4e" containerID="b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa" exitCode=0 Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.146124 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerDied","Data":"b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa"} Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.206688 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tddgt" podStartSLOduration=3.793240248 podStartE2EDuration="1m53.206664861s" podCreationTimestamp="2025-11-27 11:27:36 +0000 UTC" firstStartedPulling="2025-11-27 11:27:38.283666337 +0000 UTC m=+175.801985255" lastFinishedPulling="2025-11-27 11:29:27.69709095 +0000 UTC m=+285.215409868" observedRunningTime="2025-11-27 11:29:29.193592328 +0000 UTC m=+286.711911256" watchObservedRunningTime="2025-11-27 11:29:29.206664861 +0000 UTC m=+286.724983779" Nov 27 11:29:29 crc kubenswrapper[4796]: I1127 11:29:29.259303 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-klnx9" podStartSLOduration=3.875159629 podStartE2EDuration="1m53.259286087s" podCreationTimestamp="2025-11-27 11:27:36 +0000 UTC" firstStartedPulling="2025-11-27 11:27:38.316469899 +0000 UTC m=+175.834788817" lastFinishedPulling="2025-11-27 11:29:27.700596357 +0000 UTC m=+285.218915275" observedRunningTime="2025-11-27 11:29:29.255670396 +0000 UTC m=+286.773989364" watchObservedRunningTime="2025-11-27 11:29:29.259286087 +0000 UTC m=+286.777605015" Nov 27 11:29:37 crc kubenswrapper[4796]: I1127 11:29:37.069784 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:29:37 crc kubenswrapper[4796]: I1127 11:29:37.070594 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:29:37 crc kubenswrapper[4796]: I1127 11:29:37.246470 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:29:37 crc kubenswrapper[4796]: I1127 11:29:37.246534 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:29:38 crc kubenswrapper[4796]: I1127 11:29:38.451583 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:29:38 crc kubenswrapper[4796]: I1127 11:29:38.453572 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:29:38 crc kubenswrapper[4796]: I1127 11:29:38.526472 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:29:38 crc kubenswrapper[4796]: I1127 11:29:38.533596 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:29:39 crc kubenswrapper[4796]: I1127 11:29:39.818394 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klnx9"] Nov 27 11:29:40 crc kubenswrapper[4796]: I1127 11:29:40.229305 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-klnx9" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="registry-server" containerID="cri-o://eb49f5ec329482b9e62a2b40279bc18f02700a2762f9524eb9c0f1cc8e53478b" gracePeriod=2 Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.990372 4796 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.990754 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ce83c0-1f80-485d-970b-51beaab3f6df" containerName="pruner" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.990778 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ce83c0-1f80-485d-970b-51beaab3f6df" containerName="pruner" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.990954 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5ce83c0-1f80-485d-970b-51beaab3f6df" containerName="pruner" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.991463 4796 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.991613 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.991892 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764" gracePeriod=15 Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.991930 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd" gracePeriod=15 Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.992016 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649" gracePeriod=15 Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.992077 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91" gracePeriod=15 Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.992066 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b" gracePeriod=15 Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995081 4796 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995503 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995527 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995546 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995559 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995581 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995593 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995610 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995623 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995636 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995648 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995669 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995681 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.995704 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995716 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995940 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995967 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.995983 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.996001 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.996027 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.996040 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 11:29:41 crc kubenswrapper[4796]: E1127 11:29:41.996216 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.996232 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:41 crc kubenswrapper[4796]: I1127 11:29:41.996448 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.039343 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130004 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130106 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130258 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130331 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130360 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130396 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.130686 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231627 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231692 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231726 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231764 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231808 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231826 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231892 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231950 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231958 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.232001 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.232025 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.232061 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231938 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.231975 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.232033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.336328 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.493546 4796 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Nov 27 11:29:42 crc kubenswrapper[4796]: I1127 11:29:42.493624 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.255393 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.257945 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.259093 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b" exitCode=0 Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.259153 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd" exitCode=0 Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.259176 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91" exitCode=2 Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.259318 4796 scope.go:117] "RemoveContainer" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.263593 4796 generic.go:334] "Generic (PLEG): container finished" podID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerID="eb49f5ec329482b9e62a2b40279bc18f02700a2762f9524eb9c0f1cc8e53478b" exitCode=0 Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.263644 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerDied","Data":"eb49f5ec329482b9e62a2b40279bc18f02700a2762f9524eb9c0f1cc8e53478b"} Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.574102 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:43 crc kubenswrapper[4796]: I1127 11:29:43.574819 4796 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:44 crc kubenswrapper[4796]: I1127 11:29:44.291364 4796 generic.go:334] "Generic (PLEG): container finished" podID="99b0a522-adc4-4078-ad97-941f1597a69c" containerID="4b6cc76710e99ed37095732fc5cefa9cbcdb99521399bf3943abd959301e845d" exitCode=0 Nov 27 11:29:44 crc kubenswrapper[4796]: I1127 11:29:44.291517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"99b0a522-adc4-4078-ad97-941f1597a69c","Type":"ContainerDied","Data":"4b6cc76710e99ed37095732fc5cefa9cbcdb99521399bf3943abd959301e845d"} Nov 27 11:29:44 crc kubenswrapper[4796]: I1127 11:29:44.295512 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:44 crc kubenswrapper[4796]: I1127 11:29:44.296354 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.316112 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.318231 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649" exitCode=0 Nov 27 11:29:45 crc kubenswrapper[4796]: E1127 11:29:45.712060 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:29:45Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:29:45Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:29:45Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T11:29:45Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: E1127 11:29:45.712521 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: E1127 11:29:45.712923 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: E1127 11:29:45.713535 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: E1127 11:29:45.713960 4796 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: E1127 11:29:45.714002 4796 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.982346 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.983313 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.983784 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.990216 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.990776 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.991126 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:45 crc kubenswrapper[4796]: I1127 11:29:45.991458 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.020485 4796 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.021522 4796 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.022223 4796 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.022797 4796 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.023168 4796 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.023199 4796 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.023520 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="200ms" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.089656 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-var-lock\") pod \"99b0a522-adc4-4078-ad97-941f1597a69c\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.089755 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-var-lock" (OuterVolumeSpecName: "var-lock") pod "99b0a522-adc4-4078-ad97-941f1597a69c" (UID: "99b0a522-adc4-4078-ad97-941f1597a69c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.089830 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99b0a522-adc4-4078-ad97-941f1597a69c-kube-api-access\") pod \"99b0a522-adc4-4078-ad97-941f1597a69c\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.090184 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-catalog-content\") pod \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.090363 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-kubelet-dir\") pod \"99b0a522-adc4-4078-ad97-941f1597a69c\" (UID: \"99b0a522-adc4-4078-ad97-941f1597a69c\") " Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.090405 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "99b0a522-adc4-4078-ad97-941f1597a69c" (UID: "99b0a522-adc4-4078-ad97-941f1597a69c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.090482 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmvp7\" (UniqueName: \"kubernetes.io/projected/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-kube-api-access-qmvp7\") pod \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.090580 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-utilities\") pod \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\" (UID: \"ddf010b4-d216-4ab7-bc54-a8879fcb45e0\") " Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.091203 4796 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.091239 4796 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99b0a522-adc4-4078-ad97-941f1597a69c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.091864 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-utilities" (OuterVolumeSpecName: "utilities") pod "ddf010b4-d216-4ab7-bc54-a8879fcb45e0" (UID: "ddf010b4-d216-4ab7-bc54-a8879fcb45e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.098463 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b0a522-adc4-4078-ad97-941f1597a69c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "99b0a522-adc4-4078-ad97-941f1597a69c" (UID: "99b0a522-adc4-4078-ad97-941f1597a69c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.098718 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-kube-api-access-qmvp7" (OuterVolumeSpecName: "kube-api-access-qmvp7") pod "ddf010b4-d216-4ab7-bc54-a8879fcb45e0" (UID: "ddf010b4-d216-4ab7-bc54-a8879fcb45e0"). InnerVolumeSpecName "kube-api-access-qmvp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.195085 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmvp7\" (UniqueName: \"kubernetes.io/projected/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-kube-api-access-qmvp7\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.195122 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.195132 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99b0a522-adc4-4078-ad97-941f1597a69c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.225113 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="400ms" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.327566 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnx9" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.327570 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnx9" event={"ID":"ddf010b4-d216-4ab7-bc54-a8879fcb45e0","Type":"ContainerDied","Data":"bf761285ddd3db7ff3e3e0d329e91507b5ddd69eb398b57a503ea1332a6e741b"} Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.328950 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.329635 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.330123 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.330840 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.330863 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"99b0a522-adc4-4078-ad97-941f1597a69c","Type":"ContainerDied","Data":"cacadfa6ac3d70549e14d7249bc08982fddb61c79832060db2939c930cdc9c9b"} Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.330941 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cacadfa6ac3d70549e14d7249bc08982fddb61c79832060db2939c930cdc9c9b" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.337710 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.339028 4796 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764" exitCode=0 Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.344903 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.345363 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: I1127 11:29:46.345921 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:46 crc kubenswrapper[4796]: E1127 11:29:46.626121 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="800ms" Nov 27 11:29:47 crc kubenswrapper[4796]: E1127 11:29:47.426990 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="1.6s" Nov 27 11:29:49 crc kubenswrapper[4796]: E1127 11:29:49.028712 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="3.2s" Nov 27 11:29:52 crc kubenswrapper[4796]: E1127 11:29:52.229867 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="6.4s" Nov 27 11:29:52 crc kubenswrapper[4796]: I1127 11:29:52.950548 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 11:29:52 crc kubenswrapper[4796]: I1127 11:29:52.952111 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:52 crc kubenswrapper[4796]: I1127 11:29:52.952848 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:52 crc kubenswrapper[4796]: I1127 11:29:52.953239 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:52 crc kubenswrapper[4796]: I1127 11:29:52.953833 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:52 crc kubenswrapper[4796]: I1127 11:29:52.954491 4796 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.020366 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.020541 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.020647 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.020856 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.020970 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.020972 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.021446 4796 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.021479 4796 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.021497 4796 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.410466 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.412164 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.431053 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.431647 4796 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.433129 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.433610 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.572357 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.572708 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.572999 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.573379 4796 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.575462 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.620581 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddf010b4-d216-4ab7-bc54-a8879fcb45e0" (UID: "ddf010b4-d216-4ab7-bc54-a8879fcb45e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.631425 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddf010b4-d216-4ab7-bc54-a8879fcb45e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.845251 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.845620 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:53 crc kubenswrapper[4796]: I1127 11:29:53.845870 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.568624 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.569804 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.570212 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.570807 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.585028 4796 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.585076 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:29:54 crc kubenswrapper[4796]: E1127 11:29:54.585577 4796 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:54 crc kubenswrapper[4796]: I1127 11:29:54.586187 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.439959 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.440017 4796 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83" exitCode=1 Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.440050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83"} Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.440735 4796 scope.go:117] "RemoveContainer" containerID="907b821e4c302a6d55252cf7f9d761c83d9099cd4736667e1bbfc820ab143a83" Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.441081 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.441932 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.442580 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:56 crc kubenswrapper[4796]: I1127 11:29:56.443002 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:29:57 crc kubenswrapper[4796]: I1127 11:29:57.257866 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:29:57 crc kubenswrapper[4796]: E1127 11:29:57.362758 4796 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.145:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-92gc9.187bd9a8a33c1b39 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-92gc9,UID:906ed244-7663-473c-8495-fe03bd4aa745,APIVersion:v1,ResourceVersion:28712,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 28.23s (28.23s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,LastTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 11:29:57 crc kubenswrapper[4796]: I1127 11:29:57.608379 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:29:58 crc kubenswrapper[4796]: E1127 11:29:58.632049 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="7s" Nov 27 11:30:00 crc kubenswrapper[4796]: I1127 11:30:00.566220 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:30:02 crc kubenswrapper[4796]: I1127 11:30:02.291969 4796 scope.go:117] "RemoveContainer" containerID="eb49f5ec329482b9e62a2b40279bc18f02700a2762f9524eb9c0f1cc8e53478b" Nov 27 11:30:02 crc kubenswrapper[4796]: I1127 11:30:02.489856 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.579926 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.580578 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.581015 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.581565 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.582438 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: E1127 11:30:03.805120 4796 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.145:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-92gc9.187bd9a8a33c1b39 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-92gc9,UID:906ed244-7663-473c-8495-fe03bd4aa745,APIVersion:v1,ResourceVersion:28712,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 28.23s (28.23s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,LastTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.947010 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.947984 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.948615 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.949810 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.950230 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.950892 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:03 crc kubenswrapper[4796]: I1127 11:30:03.951363 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:05 crc kubenswrapper[4796]: E1127 11:30:05.633077 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="7s" Nov 27 11:30:12 crc kubenswrapper[4796]: E1127 11:30:12.634375 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="7s" Nov 27 11:30:13 crc kubenswrapper[4796]: I1127 11:30:13.588317 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:13 crc kubenswrapper[4796]: I1127 11:30:13.588886 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:13 crc kubenswrapper[4796]: I1127 11:30:13.589678 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:13 crc kubenswrapper[4796]: I1127 11:30:13.590346 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:13 crc kubenswrapper[4796]: I1127 11:30:13.590752 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:13 crc kubenswrapper[4796]: I1127 11:30:13.591304 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:13 crc kubenswrapper[4796]: E1127 11:30:13.806087 4796 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.145:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-92gc9.187bd9a8a33c1b39 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-92gc9,UID:906ed244-7663-473c-8495-fe03bd4aa745,APIVersion:v1,ResourceVersion:28712,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 28.23s (28.23s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,LastTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 11:30:15 crc kubenswrapper[4796]: E1127 11:30:15.574374 4796 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.145:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" volumeName="registry-storage" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.080402 4796 scope.go:117] "RemoveContainer" containerID="4a81f02b197ad0b7c58a8e561ddba5107885352bb1a5a1b8b6b877e8cd291ec7" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.597779 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-vrzqb_ef543e1b-8068-4ea3-b32a-61027b32e95d/approver/0.log" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.598987 4796 generic.go:334] "Generic (PLEG): container finished" podID="ef543e1b-8068-4ea3-b32a-61027b32e95d" containerID="fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db" exitCode=1 Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.599101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerDied","Data":"fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db"} Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.600236 4796 scope.go:117] "RemoveContainer" containerID="fa4e03a8dd624736087a6d551df8d4e5c1925481dcda49d5ec0afd2282ea21db" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.601392 4796 status_manager.go:851] "Failed to get status for pod" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/pods/network-node-identity-vrzqb\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.602068 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.603237 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.603818 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.604590 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.605235 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:18 crc kubenswrapper[4796]: I1127 11:30:18.606425 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:19 crc kubenswrapper[4796]: E1127 11:30:19.635898 4796 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.145:6443: connect: connection refused" interval="7s" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.087427 4796 scope.go:117] "RemoveContainer" containerID="13ff7d0016887d4175a933baa62bc6d90d6469f1ddf19863671dce6d8c9eea7a" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.190678 4796 scope.go:117] "RemoveContainer" containerID="c16bc0449ab6c390ba98ac6314eb439fd8f9c93ee19cd2c8a7b648addb0c391b" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.234777 4796 scope.go:117] "RemoveContainer" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:30:20 crc kubenswrapper[4796]: E1127 11:30:20.235580 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\": container with ID starting with 711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1 not found: ID does not exist" containerID="711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.235618 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1"} err="failed to get container status \"711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\": rpc error: code = NotFound desc = could not find container \"711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1\": container with ID starting with 711f602a3c54e40864d9ced05a5d719993ca932d42e1cf0a85e0448afdde6bd1 not found: ID does not exist" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.235642 4796 scope.go:117] "RemoveContainer" containerID="8b0071da5384187247a8b31a193eb48923c210a7da42313d1854d6909a1bf5dd" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.280886 4796 scope.go:117] "RemoveContainer" containerID="c1742fd229328ddba8a563f65bb76f689d914fe1a7093aa584f53a3f1a721649" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.302698 4796 scope.go:117] "RemoveContainer" containerID="e2b75903f6b5edb7d3c11853298ef8780239d890430079d7282b6fc443dbae91" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.319804 4796 scope.go:117] "RemoveContainer" containerID="6b7f83e0227ad0b4239a838d0dc8e8a3b525a3252de19c2ce46a50f8d6604764" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.353187 4796 scope.go:117] "RemoveContainer" containerID="9c308af6da21c1cab43e7930b2cbbfca5327b5628388256d312c5da21c3a0964" Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.626633 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d21dec2fe38a77be3f3ca0b87736c1fe0380626e09123729294d6c4c3384b464"} Nov 27 11:30:20 crc kubenswrapper[4796]: I1127 11:30:20.632385 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"9a40cb1b7826a801b27e75127b2d179083c8e447c2901bc0b6c60d1ec64d394d"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.647547 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"94f57b3f58007a96eb1c540d7cbb4ce663c36d2c373f607f69d6702a23982596"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.650407 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerStarted","Data":"de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.651155 4796 status_manager.go:851] "Failed to get status for pod" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/pods/network-node-identity-vrzqb\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.651540 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.652130 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.652510 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.652686 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.652784 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerStarted","Data":"78b69ab0ef2d4ffb9a2e80e800e71a8ed1171353975e169b9ff5084834637056"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.652901 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.653154 4796 status_manager.go:851] "Failed to get status for pod" podUID="906ed244-7663-473c-8495-fe03bd4aa745" pod="openshift-marketplace/redhat-operators-92gc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-92gc9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.653356 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.654965 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerStarted","Data":"ad93c7a7ffa992ef8e13a58a0921b45dde896687c1fb80467250d073206d70c5"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.657078 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-vrzqb_ef543e1b-8068-4ea3-b32a-61027b32e95d/approver/0.log" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.657594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d0c7577b47cb90852d5022bc13e2b178fcf76b50bd325e978f43a44e0b4e54f7"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.663162 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerStarted","Data":"34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.664295 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"065cbae7651d0076830822e3631e3849639f0bf125678880ac9bbf4722302231"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.666096 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerStarted","Data":"c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.668583 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.668646 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"729697a1c6ab5de70f58284b99cc5cae88dee0f52de235698e6b06af51db5997"} Nov 27 11:30:22 crc kubenswrapper[4796]: I1127 11:30:22.670436 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerStarted","Data":"5bbd3707748056bdc1110499fbfd78cfdabfd6e0ed331fc2a5a742d0a2b88b1b"} Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.572640 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.573362 4796 status_manager.go:851] "Failed to get status for pod" podUID="906ed244-7663-473c-8495-fe03bd4aa745" pod="openshift-marketplace/redhat-operators-92gc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-92gc9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.573656 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.573945 4796 status_manager.go:851] "Failed to get status for pod" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/pods/network-node-identity-vrzqb\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.574236 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.574685 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.574978 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.575220 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.676372 4796 generic.go:334] "Generic (PLEG): container finished" podID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerID="78b69ab0ef2d4ffb9a2e80e800e71a8ed1171353975e169b9ff5084834637056" exitCode=0 Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.676452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerDied","Data":"78b69ab0ef2d4ffb9a2e80e800e71a8ed1171353975e169b9ff5084834637056"} Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.677085 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.677507 4796 status_manager.go:851] "Failed to get status for pod" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/pods/network-node-identity-vrzqb\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.677832 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.678002 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.678181 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.678387 4796 status_manager.go:851] "Failed to get status for pod" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" pod="openshift-marketplace/redhat-marketplace-djbsd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-djbsd\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.678572 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.678773 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.678954 4796 status_manager.go:851] "Failed to get status for pod" podUID="906ed244-7663-473c-8495-fe03bd4aa745" pod="openshift-marketplace/redhat-operators-92gc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-92gc9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.679070 4796 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="065cbae7651d0076830822e3631e3849639f0bf125678880ac9bbf4722302231" exitCode=0 Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.679382 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"065cbae7651d0076830822e3631e3849639f0bf125678880ac9bbf4722302231"} Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.679701 4796 status_manager.go:851] "Failed to get status for pod" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/pods/network-node-identity-vrzqb\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.679886 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.679958 4796 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.679970 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.680051 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: E1127 11:30:23.680137 4796 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.680235 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.681666 4796 status_manager.go:851] "Failed to get status for pod" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" pod="openshift-marketplace/redhat-marketplace-djbsd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-djbsd\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.681964 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.682168 4796 status_manager.go:851] "Failed to get status for pod" podUID="906ed244-7663-473c-8495-fe03bd4aa745" pod="openshift-marketplace/redhat-operators-92gc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-92gc9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.682408 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.682694 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.682969 4796 status_manager.go:851] "Failed to get status for pod" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" pod="openshift-marketplace/redhat-operators-2j2sz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2j2sz\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.683217 4796 status_manager.go:851] "Failed to get status for pod" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" pod="openshift-marketplace/certified-operators-hrwn6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-hrwn6\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.683455 4796 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.683757 4796 status_manager.go:851] "Failed to get status for pod" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" pod="openshift-marketplace/redhat-marketplace-djbsd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-djbsd\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.684082 4796 status_manager.go:851] "Failed to get status for pod" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" pod="openshift-marketplace/community-operators-klnx9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-klnx9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.684337 4796 status_manager.go:851] "Failed to get status for pod" podUID="906ed244-7663-473c-8495-fe03bd4aa745" pod="openshift-marketplace/redhat-operators-92gc9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-92gc9\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.684537 4796 status_manager.go:851] "Failed to get status for pod" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" pod="openshift-marketplace/redhat-marketplace-pbl4f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-pbl4f\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.684733 4796 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.684925 4796 status_manager.go:851] "Failed to get status for pod" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" pod="openshift-marketplace/redhat-operators-2j2sz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2j2sz\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.685118 4796 status_manager.go:851] "Failed to get status for pod" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" pod="openshift-marketplace/community-operators-hxkmw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-hxkmw\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.685354 4796 status_manager.go:851] "Failed to get status for pod" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/pods/network-node-identity-vrzqb\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.685551 4796 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.685741 4796 status_manager.go:851] "Failed to get status for pod" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: I1127 11:30:23.685927 4796 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.145:6443: connect: connection refused" Nov 27 11:30:23 crc kubenswrapper[4796]: E1127 11:30:23.806909 4796 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.145:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-92gc9.187bd9a8a33c1b39 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-92gc9,UID:906ed244-7663-473c-8495-fe03bd4aa745,APIVersion:v1,ResourceVersion:28712,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 28.23s (28.23s including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,LastTimestamp:2025-11-27 11:29:57.362039609 +0000 UTC m=+314.880358567,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 11:30:24 crc kubenswrapper[4796]: I1127 11:30:24.689808 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a9005acb9348745ffa91ff7e925c68d08badf35485013509fd58bea996383897"} Nov 27 11:30:24 crc kubenswrapper[4796]: I1127 11:30:24.690348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c2b6654fdcc2c0e1ebea9832f3d7cae878c524e9d12c77f91397b787470751a9"} Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.698333 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"21069b7c518bd9ecd45cd5105a96fe555587505b2c01e9e9e3d8c74979502246"} Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.698909 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"655151a55855080c0a9c9d1fa097fc01ef098c0a6f7e05ab166fc33a676010c9"} Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.698926 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ac6446b2ab39dc4d8907da61fce8059d97f661639714484f1c694d91c43e94c4"} Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.699204 4796 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.699222 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.699440 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:25 crc kubenswrapper[4796]: I1127 11:30:25.701814 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerStarted","Data":"b3709bc2f8a23ebe4c9a8af046625474a487a3e20ed34df4d9975cfd4434740a"} Nov 27 11:30:26 crc kubenswrapper[4796]: I1127 11:30:26.884686 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:30:26 crc kubenswrapper[4796]: I1127 11:30:26.884960 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:30:26 crc kubenswrapper[4796]: I1127 11:30:26.930444 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.608403 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.614373 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.711690 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.759642 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.760000 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.760076 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:30:27 crc kubenswrapper[4796]: I1127 11:30:27.802927 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:30:28 crc kubenswrapper[4796]: I1127 11:30:28.755383 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.046236 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.046308 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.088844 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.488844 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.488894 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.527116 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.587123 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.587173 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.592446 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.755095 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:30:29 crc kubenswrapper[4796]: I1127 11:30:29.763198 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.077663 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.077738 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.126492 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.302947 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.303038 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.348750 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.714687 4796 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.771845 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.771959 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:30:30 crc kubenswrapper[4796]: I1127 11:30:30.779928 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f1dfc794-4ffa-4208-95cc-d1a414dadf17" Nov 27 11:30:31 crc kubenswrapper[4796]: I1127 11:30:31.730653 4796 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:31 crc kubenswrapper[4796]: I1127 11:30:31.730682 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:31 crc kubenswrapper[4796]: I1127 11:30:31.733806 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f1dfc794-4ffa-4208-95cc-d1a414dadf17" Nov 27 11:30:31 crc kubenswrapper[4796]: I1127 11:30:31.734997 4796 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://c2b6654fdcc2c0e1ebea9832f3d7cae878c524e9d12c77f91397b787470751a9" Nov 27 11:30:31 crc kubenswrapper[4796]: I1127 11:30:31.735028 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:30:32 crc kubenswrapper[4796]: I1127 11:30:32.734938 4796 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:32 crc kubenswrapper[4796]: I1127 11:30:32.734971 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:30:32 crc kubenswrapper[4796]: I1127 11:30:32.737633 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f1dfc794-4ffa-4208-95cc-d1a414dadf17" Nov 27 11:30:40 crc kubenswrapper[4796]: I1127 11:30:40.570329 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 11:30:56 crc kubenswrapper[4796]: I1127 11:30:56.367917 4796 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 11:30:56 crc kubenswrapper[4796]: I1127 11:30:56.410309 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 11:30:56 crc kubenswrapper[4796]: I1127 11:30:56.647753 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 11:30:56 crc kubenswrapper[4796]: I1127 11:30:56.991969 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 11:30:57 crc kubenswrapper[4796]: I1127 11:30:57.196357 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 11:30:57 crc kubenswrapper[4796]: I1127 11:30:57.305748 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 11:30:57 crc kubenswrapper[4796]: I1127 11:30:57.311915 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 11:30:57 crc kubenswrapper[4796]: I1127 11:30:57.462814 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 11:30:57 crc kubenswrapper[4796]: I1127 11:30:57.799376 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 11:30:58 crc kubenswrapper[4796]: I1127 11:30:58.491440 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 11:30:58 crc kubenswrapper[4796]: I1127 11:30:58.544399 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 11:30:58 crc kubenswrapper[4796]: I1127 11:30:58.703536 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 11:30:58 crc kubenswrapper[4796]: I1127 11:30:58.907968 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 11:30:59 crc kubenswrapper[4796]: I1127 11:30:59.563010 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 11:31:00 crc kubenswrapper[4796]: I1127 11:31:00.308554 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 11:31:00 crc kubenswrapper[4796]: I1127 11:31:00.309427 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 11:31:00 crc kubenswrapper[4796]: I1127 11:31:00.499218 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 11:31:01 crc kubenswrapper[4796]: I1127 11:31:01.539119 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 11:31:02 crc kubenswrapper[4796]: I1127 11:31:02.265746 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 11:31:02 crc kubenswrapper[4796]: I1127 11:31:02.791306 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.428976 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.670716 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.695612 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.770799 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.800991 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.938235 4796 generic.go:334] "Generic (PLEG): container finished" podID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerID="eab8c03db19581f6cd9186955c3d21d66771e944336a1df0454ef7c0a4e6d96e" exitCode=0 Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.938292 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerDied","Data":"eab8c03db19581f6cd9186955c3d21d66771e944336a1df0454ef7c0a4e6d96e"} Nov 27 11:31:03 crc kubenswrapper[4796]: I1127 11:31:03.938743 4796 scope.go:117] "RemoveContainer" containerID="eab8c03db19581f6cd9186955c3d21d66771e944336a1df0454ef7c0a4e6d96e" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.025539 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.137650 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.201358 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.740484 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.947626 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fdwzj_dc437d52-8b8b-430f-bf9e-67492c365aad/marketplace-operator/1.log" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.949038 4796 generic.go:334] "Generic (PLEG): container finished" podID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" exitCode=1 Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.949083 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerDied","Data":"4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420"} Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.949149 4796 scope.go:117] "RemoveContainer" containerID="eab8c03db19581f6cd9186955c3d21d66771e944336a1df0454ef7c0a4e6d96e" Nov 27 11:31:04 crc kubenswrapper[4796]: I1127 11:31:04.949688 4796 scope.go:117] "RemoveContainer" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" Nov 27 11:31:04 crc kubenswrapper[4796]: E1127 11:31:04.950084 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-fdwzj_openshift-marketplace(dc437d52-8b8b-430f-bf9e-67492c365aad)\"" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" Nov 27 11:31:05 crc kubenswrapper[4796]: I1127 11:31:05.722395 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 11:31:05 crc kubenswrapper[4796]: I1127 11:31:05.768921 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:31:05 crc kubenswrapper[4796]: I1127 11:31:05.768999 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:31:05 crc kubenswrapper[4796]: I1127 11:31:05.830367 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 11:31:05 crc kubenswrapper[4796]: I1127 11:31:05.958082 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fdwzj_dc437d52-8b8b-430f-bf9e-67492c365aad/marketplace-operator/1.log" Nov 27 11:31:05 crc kubenswrapper[4796]: I1127 11:31:05.958516 4796 scope.go:117] "RemoveContainer" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" Nov 27 11:31:05 crc kubenswrapper[4796]: E1127 11:31:05.958700 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-fdwzj_openshift-marketplace(dc437d52-8b8b-430f-bf9e-67492c365aad)\"" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" Nov 27 11:31:06 crc kubenswrapper[4796]: I1127 11:31:06.001405 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 11:31:07 crc kubenswrapper[4796]: I1127 11:31:07.339013 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 11:31:08 crc kubenswrapper[4796]: I1127 11:31:08.084153 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 11:31:08 crc kubenswrapper[4796]: I1127 11:31:08.618430 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 11:31:08 crc kubenswrapper[4796]: I1127 11:31:08.846468 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 11:31:08 crc kubenswrapper[4796]: I1127 11:31:08.861590 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 11:31:09 crc kubenswrapper[4796]: I1127 11:31:09.068593 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 11:31:09 crc kubenswrapper[4796]: I1127 11:31:09.115758 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 11:31:09 crc kubenswrapper[4796]: I1127 11:31:09.300217 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 11:31:09 crc kubenswrapper[4796]: I1127 11:31:09.607251 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 11:31:09 crc kubenswrapper[4796]: I1127 11:31:09.692580 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 11:31:10 crc kubenswrapper[4796]: I1127 11:31:10.053875 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 11:31:10 crc kubenswrapper[4796]: I1127 11:31:10.233430 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 11:31:10 crc kubenswrapper[4796]: I1127 11:31:10.574777 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 11:31:11 crc kubenswrapper[4796]: I1127 11:31:11.466340 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 11:31:12 crc kubenswrapper[4796]: I1127 11:31:12.604738 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 11:31:12 crc kubenswrapper[4796]: I1127 11:31:12.870352 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 11:31:13 crc kubenswrapper[4796]: I1127 11:31:13.076480 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 11:31:13 crc kubenswrapper[4796]: I1127 11:31:13.448127 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 11:31:13 crc kubenswrapper[4796]: I1127 11:31:13.835729 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 11:31:13 crc kubenswrapper[4796]: I1127 11:31:13.879209 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 11:31:14 crc kubenswrapper[4796]: I1127 11:31:14.067839 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 11:31:14 crc kubenswrapper[4796]: I1127 11:31:14.301582 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 11:31:15 crc kubenswrapper[4796]: I1127 11:31:15.530381 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 11:31:15 crc kubenswrapper[4796]: I1127 11:31:15.601566 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 11:31:15 crc kubenswrapper[4796]: I1127 11:31:15.693392 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 11:31:16 crc kubenswrapper[4796]: I1127 11:31:16.653969 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 11:31:16 crc kubenswrapper[4796]: I1127 11:31:16.731125 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 11:31:16 crc kubenswrapper[4796]: I1127 11:31:16.908747 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.005217 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.070897 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.506914 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.627671 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.720029 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.734089 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.818732 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 11:31:17 crc kubenswrapper[4796]: I1127 11:31:17.820572 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 11:31:18 crc kubenswrapper[4796]: I1127 11:31:18.067912 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 11:31:18 crc kubenswrapper[4796]: I1127 11:31:18.139335 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 11:31:18 crc kubenswrapper[4796]: I1127 11:31:18.187977 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 11:31:18 crc kubenswrapper[4796]: I1127 11:31:18.480429 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 11:31:18 crc kubenswrapper[4796]: I1127 11:31:18.962954 4796 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 11:31:19 crc kubenswrapper[4796]: I1127 11:31:19.352054 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 11:31:19 crc kubenswrapper[4796]: I1127 11:31:19.733749 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 11:31:20 crc kubenswrapper[4796]: I1127 11:31:20.491547 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 11:31:20 crc kubenswrapper[4796]: I1127 11:31:20.989776 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 11:31:21 crc kubenswrapper[4796]: I1127 11:31:21.488076 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 11:31:21 crc kubenswrapper[4796]: I1127 11:31:21.569480 4796 scope.go:117] "RemoveContainer" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" Nov 27 11:31:21 crc kubenswrapper[4796]: I1127 11:31:21.643826 4796 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 11:31:22 crc kubenswrapper[4796]: I1127 11:31:22.079400 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fdwzj_dc437d52-8b8b-430f-bf9e-67492c365aad/marketplace-operator/1.log" Nov 27 11:31:22 crc kubenswrapper[4796]: I1127 11:31:22.079465 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerStarted","Data":"e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d"} Nov 27 11:31:22 crc kubenswrapper[4796]: I1127 11:31:22.079794 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:31:22 crc kubenswrapper[4796]: I1127 11:31:22.082000 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:31:22 crc kubenswrapper[4796]: I1127 11:31:22.782308 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 11:31:22 crc kubenswrapper[4796]: I1127 11:31:22.896689 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 11:31:23 crc kubenswrapper[4796]: I1127 11:31:23.082032 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 11:31:23 crc kubenswrapper[4796]: I1127 11:31:23.630497 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 11:31:24 crc kubenswrapper[4796]: I1127 11:31:24.128754 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 11:31:24 crc kubenswrapper[4796]: I1127 11:31:24.268586 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 11:31:24 crc kubenswrapper[4796]: I1127 11:31:24.798894 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 11:31:24 crc kubenswrapper[4796]: I1127 11:31:24.957956 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 11:31:25 crc kubenswrapper[4796]: I1127 11:31:25.575917 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 11:31:25 crc kubenswrapper[4796]: I1127 11:31:25.799857 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 11:31:26 crc kubenswrapper[4796]: I1127 11:31:26.779064 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 11:31:26 crc kubenswrapper[4796]: I1127 11:31:26.993923 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 11:31:27 crc kubenswrapper[4796]: I1127 11:31:27.366195 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 11:31:27 crc kubenswrapper[4796]: I1127 11:31:27.953676 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 11:31:27 crc kubenswrapper[4796]: I1127 11:31:27.955969 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 11:31:27 crc kubenswrapper[4796]: I1127 11:31:27.976976 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 11:31:28 crc kubenswrapper[4796]: I1127 11:31:28.085592 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 11:31:28 crc kubenswrapper[4796]: I1127 11:31:28.358089 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 11:31:28 crc kubenswrapper[4796]: I1127 11:31:28.788077 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 11:31:28 crc kubenswrapper[4796]: I1127 11:31:28.865010 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 11:31:29 crc kubenswrapper[4796]: I1127 11:31:29.068038 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 11:31:29 crc kubenswrapper[4796]: I1127 11:31:29.705321 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 11:31:29 crc kubenswrapper[4796]: I1127 11:31:29.835883 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 11:31:29 crc kubenswrapper[4796]: I1127 11:31:29.896461 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 11:31:29 crc kubenswrapper[4796]: I1127 11:31:29.908229 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 11:31:29 crc kubenswrapper[4796]: I1127 11:31:29.946461 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 11:31:30 crc kubenswrapper[4796]: I1127 11:31:30.877975 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.016080 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.145944 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.242309 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.523874 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.566886 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.826121 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.837291 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.879832 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:31:31 crc kubenswrapper[4796]: I1127 11:31:31.879945 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:31:32 crc kubenswrapper[4796]: I1127 11:31:32.365810 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 11:31:32 crc kubenswrapper[4796]: I1127 11:31:32.783922 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 11:31:32 crc kubenswrapper[4796]: I1127 11:31:32.878865 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 11:31:32 crc kubenswrapper[4796]: I1127 11:31:32.987562 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 11:31:33 crc kubenswrapper[4796]: I1127 11:31:33.246140 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 11:31:33 crc kubenswrapper[4796]: I1127 11:31:33.303968 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 11:31:33 crc kubenswrapper[4796]: I1127 11:31:33.463092 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 11:31:33 crc kubenswrapper[4796]: I1127 11:31:33.518184 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 11:31:33 crc kubenswrapper[4796]: I1127 11:31:33.716348 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 11:31:34 crc kubenswrapper[4796]: I1127 11:31:34.133918 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 11:31:34 crc kubenswrapper[4796]: I1127 11:31:34.161357 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 11:31:34 crc kubenswrapper[4796]: I1127 11:31:34.534864 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 11:31:34 crc kubenswrapper[4796]: I1127 11:31:34.895005 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 11:31:35 crc kubenswrapper[4796]: I1127 11:31:35.055673 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 11:31:35 crc kubenswrapper[4796]: I1127 11:31:35.596197 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 11:31:35 crc kubenswrapper[4796]: I1127 11:31:35.676095 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 11:31:35 crc kubenswrapper[4796]: I1127 11:31:35.768949 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 11:31:36 crc kubenswrapper[4796]: I1127 11:31:36.158122 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 11:31:36 crc kubenswrapper[4796]: I1127 11:31:36.224865 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 11:31:36 crc kubenswrapper[4796]: I1127 11:31:36.375141 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 11:31:36 crc kubenswrapper[4796]: I1127 11:31:36.806968 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 11:31:37 crc kubenswrapper[4796]: I1127 11:31:37.333886 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 11:31:37 crc kubenswrapper[4796]: I1127 11:31:37.682093 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.420668 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.616371 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.688589 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.696737 4796 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.697161 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hxkmw" podStartSLOduration=119.700059366 podStartE2EDuration="4m2.697112872s" podCreationTimestamp="2025-11-27 11:27:36 +0000 UTC" firstStartedPulling="2025-11-27 11:27:38.290827701 +0000 UTC m=+175.809146619" lastFinishedPulling="2025-11-27 11:29:41.287881177 +0000 UTC m=+298.806200125" observedRunningTime="2025-11-27 11:30:30.938884514 +0000 UTC m=+348.457203432" watchObservedRunningTime="2025-11-27 11:31:38.697112872 +0000 UTC m=+416.215431810" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.697521 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pbl4f" podStartSLOduration=84.022834521 podStartE2EDuration="4m0.697508813s" podCreationTimestamp="2025-11-27 11:27:38 +0000 UTC" firstStartedPulling="2025-11-27 11:27:41.374083837 +0000 UTC m=+178.892402755" lastFinishedPulling="2025-11-27 11:30:18.048758099 +0000 UTC m=+335.567077047" observedRunningTime="2025-11-27 11:30:30.891880685 +0000 UTC m=+348.410199603" watchObservedRunningTime="2025-11-27 11:31:38.697508813 +0000 UTC m=+416.215827781" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.697757 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-92gc9" podStartSLOduration=103.709479099 podStartE2EDuration="3m59.697747999s" podCreationTimestamp="2025-11-27 11:27:39 +0000 UTC" firstStartedPulling="2025-11-27 11:27:41.373751629 +0000 UTC m=+178.892070547" lastFinishedPulling="2025-11-27 11:29:57.362020489 +0000 UTC m=+314.880339447" observedRunningTime="2025-11-27 11:30:30.873024239 +0000 UTC m=+348.391343157" watchObservedRunningTime="2025-11-27 11:31:38.697747999 +0000 UTC m=+416.216066937" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.698306 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-djbsd" podStartSLOduration=76.301742617 podStartE2EDuration="3m59.698295974s" podCreationTimestamp="2025-11-27 11:27:39 +0000 UTC" firstStartedPulling="2025-11-27 11:27:41.35046117 +0000 UTC m=+178.868780118" lastFinishedPulling="2025-11-27 11:30:24.747014557 +0000 UTC m=+342.265333475" observedRunningTime="2025-11-27 11:30:30.995957733 +0000 UTC m=+348.514276651" watchObservedRunningTime="2025-11-27 11:31:38.698295974 +0000 UTC m=+416.216614902" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.701294 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2j2sz" podStartSLOduration=84.575357877 podStartE2EDuration="3m58.701284497s" podCreationTimestamp="2025-11-27 11:27:40 +0000 UTC" firstStartedPulling="2025-11-27 11:27:41.374475587 +0000 UTC m=+178.892794505" lastFinishedPulling="2025-11-27 11:30:15.500402167 +0000 UTC m=+333.018721125" observedRunningTime="2025-11-27 11:30:30.923860535 +0000 UTC m=+348.442179453" watchObservedRunningTime="2025-11-27 11:31:38.701284497 +0000 UTC m=+416.219603425" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.701819 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hrwn6" podStartSLOduration=96.844368572 podStartE2EDuration="4m1.701811241s" podCreationTimestamp="2025-11-27 11:27:37 +0000 UTC" firstStartedPulling="2025-11-27 11:27:39.326682498 +0000 UTC m=+176.845001426" lastFinishedPulling="2025-11-27 11:30:04.184125137 +0000 UTC m=+321.702444095" observedRunningTime="2025-11-27 11:30:30.794364949 +0000 UTC m=+348.312683867" watchObservedRunningTime="2025-11-27 11:31:38.701811241 +0000 UTC m=+416.220130179" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.703523 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=116.703513327 podStartE2EDuration="1m56.703513327s" podCreationTimestamp="2025-11-27 11:29:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:30:30.906242835 +0000 UTC m=+348.424561753" watchObservedRunningTime="2025-11-27 11:31:38.703513327 +0000 UTC m=+416.221832255" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.704085 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-klnx9","openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.704323 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj","openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.704926 4796 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.704959 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8e4b2015-2757-4d7d-8d05-7f2f0cc9f1d4" Nov 27 11:31:38 crc kubenswrapper[4796]: E1127 11:31:38.705194 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" containerName="installer" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.705316 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" containerName="installer" Nov 27 11:31:38 crc kubenswrapper[4796]: E1127 11:31:38.705424 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="extract-utilities" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.705535 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="extract-utilities" Nov 27 11:31:38 crc kubenswrapper[4796]: E1127 11:31:38.705644 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="extract-content" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.705725 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="extract-content" Nov 27 11:31:38 crc kubenswrapper[4796]: E1127 11:31:38.705807 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="registry-server" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.705886 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="registry-server" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.706103 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b0a522-adc4-4078-ad97-941f1597a69c" containerName="installer" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.706226 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" containerName="registry-server" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.708234 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.710532 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.711149 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.733924 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.73389917 podStartE2EDuration="1m8.73389917s" podCreationTimestamp="2025-11-27 11:30:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:31:38.724592056 +0000 UTC m=+416.242911004" watchObservedRunningTime="2025-11-27 11:31:38.73389917 +0000 UTC m=+416.252218108" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.746941 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.795255 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.838533 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b725a2e-30e9-455e-b359-9f0ac33eaf00-config-volume\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.838608 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qpns\" (UniqueName: \"kubernetes.io/projected/6b725a2e-30e9-455e-b359-9f0ac33eaf00-kube-api-access-9qpns\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.838656 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b725a2e-30e9-455e-b359-9f0ac33eaf00-secret-volume\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.940322 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b725a2e-30e9-455e-b359-9f0ac33eaf00-config-volume\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.940440 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qpns\" (UniqueName: \"kubernetes.io/projected/6b725a2e-30e9-455e-b359-9f0ac33eaf00-kube-api-access-9qpns\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.940544 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b725a2e-30e9-455e-b359-9f0ac33eaf00-secret-volume\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.941558 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b725a2e-30e9-455e-b359-9f0ac33eaf00-config-volume\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.947502 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.951524 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b725a2e-30e9-455e-b359-9f0ac33eaf00-secret-volume\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:38 crc kubenswrapper[4796]: I1127 11:31:38.964177 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qpns\" (UniqueName: \"kubernetes.io/projected/6b725a2e-30e9-455e-b359-9f0ac33eaf00-kube-api-access-9qpns\") pod \"collect-profiles-29404050-fzdcj\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.037223 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.126906 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.200858 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.231236 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.269349 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.336459 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.354519 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.431538 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.490105 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.577226 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf010b4-d216-4ab7-bc54-a8879fcb45e0" path="/var/lib/kubelet/pods/ddf010b4-d216-4ab7-bc54-a8879fcb45e0/volumes" Nov 27 11:31:39 crc kubenswrapper[4796]: I1127 11:31:39.736982 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.029726 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.068160 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.069156 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.252149 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.377250 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.785891 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.844793 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 11:31:40 crc kubenswrapper[4796]: I1127 11:31:40.926545 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.085743 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.183078 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.327402 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.435348 4796 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.592055 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.671904 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.827414 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 11:31:41 crc kubenswrapper[4796]: I1127 11:31:41.893241 4796 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 11:31:42 crc kubenswrapper[4796]: I1127 11:31:42.339710 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 11:31:42 crc kubenswrapper[4796]: I1127 11:31:42.457978 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 11:31:42 crc kubenswrapper[4796]: I1127 11:31:42.517303 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 11:31:43 crc kubenswrapper[4796]: I1127 11:31:43.296025 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 11:31:43 crc kubenswrapper[4796]: I1127 11:31:43.553143 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 11:31:43 crc kubenswrapper[4796]: I1127 11:31:43.592242 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 11:31:43 crc kubenswrapper[4796]: I1127 11:31:43.811523 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 11:31:43 crc kubenswrapper[4796]: I1127 11:31:43.943654 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 11:31:44 crc kubenswrapper[4796]: I1127 11:31:44.005364 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 11:31:44 crc kubenswrapper[4796]: I1127 11:31:44.179858 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 11:31:44 crc kubenswrapper[4796]: I1127 11:31:44.523830 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 11:31:44 crc kubenswrapper[4796]: I1127 11:31:44.631607 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 11:31:44 crc kubenswrapper[4796]: I1127 11:31:44.635834 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 11:31:45 crc kubenswrapper[4796]: I1127 11:31:45.404107 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 11:31:45 crc kubenswrapper[4796]: I1127 11:31:45.433765 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 11:31:45 crc kubenswrapper[4796]: I1127 11:31:45.997131 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 11:31:46 crc kubenswrapper[4796]: I1127 11:31:46.446810 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 11:31:46 crc kubenswrapper[4796]: I1127 11:31:46.475032 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 11:31:46 crc kubenswrapper[4796]: I1127 11:31:46.817208 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 11:31:46 crc kubenswrapper[4796]: I1127 11:31:46.818642 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 11:31:46 crc kubenswrapper[4796]: I1127 11:31:46.866377 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 11:31:46 crc kubenswrapper[4796]: I1127 11:31:46.959314 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 11:31:47 crc kubenswrapper[4796]: I1127 11:31:47.084618 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 11:31:47 crc kubenswrapper[4796]: I1127 11:31:47.160195 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 11:31:47 crc kubenswrapper[4796]: I1127 11:31:47.286523 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.009426 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.085328 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.202212 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.404615 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.411063 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.758004 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 11:31:48 crc kubenswrapper[4796]: I1127 11:31:48.762652 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 11:31:49 crc kubenswrapper[4796]: I1127 11:31:49.378412 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 11:31:49 crc kubenswrapper[4796]: I1127 11:31:49.767172 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 11:31:49 crc kubenswrapper[4796]: I1127 11:31:49.942979 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 11:31:50 crc kubenswrapper[4796]: I1127 11:31:50.014093 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 11:31:50 crc kubenswrapper[4796]: I1127 11:31:50.076787 4796 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 11:31:50 crc kubenswrapper[4796]: I1127 11:31:50.077680 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://94f57b3f58007a96eb1c540d7cbb4ce663c36d2c373f607f69d6702a23982596" gracePeriod=5 Nov 27 11:31:50 crc kubenswrapper[4796]: I1127 11:31:50.375676 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 11:31:50 crc kubenswrapper[4796]: I1127 11:31:50.803905 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 11:31:50 crc kubenswrapper[4796]: I1127 11:31:50.843198 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.060077 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.103001 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.160550 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.240413 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.388126 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.469021 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.619557 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 11:31:51 crc kubenswrapper[4796]: I1127 11:31:51.731953 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 11:31:52 crc kubenswrapper[4796]: I1127 11:31:52.477760 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 11:31:52 crc kubenswrapper[4796]: I1127 11:31:52.644745 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 11:31:52 crc kubenswrapper[4796]: I1127 11:31:52.975768 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 11:31:53 crc kubenswrapper[4796]: I1127 11:31:53.998860 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 11:31:54 crc kubenswrapper[4796]: I1127 11:31:54.053386 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 11:31:54 crc kubenswrapper[4796]: I1127 11:31:54.356115 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 11:31:54 crc kubenswrapper[4796]: I1127 11:31:54.365514 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 11:31:54 crc kubenswrapper[4796]: I1127 11:31:54.468708 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.042906 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.107935 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.307728 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.307899 4796 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="94f57b3f58007a96eb1c540d7cbb4ce663c36d2c373f607f69d6702a23982596" exitCode=137 Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.355220 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.396713 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.479515 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.570828 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.577092 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.659572 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.659659 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.692447 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.729815 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.782618 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.782665 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.782689 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.782749 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.782772 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.782984 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.783013 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.783444 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.783522 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.790398 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.884401 4796 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.884434 4796 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.884442 4796 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.884452 4796 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:31:55 crc kubenswrapper[4796]: I1127 11:31:55.884462 4796 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 27 11:31:56 crc kubenswrapper[4796]: I1127 11:31:56.138478 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 11:31:56 crc kubenswrapper[4796]: I1127 11:31:56.194326 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 11:31:56 crc kubenswrapper[4796]: I1127 11:31:56.314099 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 11:31:56 crc kubenswrapper[4796]: I1127 11:31:56.314560 4796 scope.go:117] "RemoveContainer" containerID="94f57b3f58007a96eb1c540d7cbb4ce663c36d2c373f607f69d6702a23982596" Nov 27 11:31:56 crc kubenswrapper[4796]: I1127 11:31:56.314608 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 11:31:56 crc kubenswrapper[4796]: I1127 11:31:56.499406 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.134862 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.579346 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.580642 4796 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.589685 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.589887 4796 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="908764b5-a009-4335-8da5-4dfaa3eb7634" Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.593292 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.593339 4796 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="908764b5-a009-4335-8da5-4dfaa3eb7634" Nov 27 11:31:57 crc kubenswrapper[4796]: I1127 11:31:57.948588 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 11:31:58 crc kubenswrapper[4796]: I1127 11:31:58.040681 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 11:31:59 crc kubenswrapper[4796]: I1127 11:31:59.112687 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj"] Nov 27 11:31:59 crc kubenswrapper[4796]: I1127 11:31:59.330048 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj"] Nov 27 11:31:59 crc kubenswrapper[4796]: I1127 11:31:59.435508 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 11:32:00 crc kubenswrapper[4796]: I1127 11:32:00.003549 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 11:32:00 crc kubenswrapper[4796]: I1127 11:32:00.339775 4796 generic.go:334] "Generic (PLEG): container finished" podID="6b725a2e-30e9-455e-b359-9f0ac33eaf00" containerID="7bd33a3aaa34f511ff2a890b3ed6ce4955f5f3e3c1d65214160e7fab39cbbb71" exitCode=0 Nov 27 11:32:00 crc kubenswrapper[4796]: I1127 11:32:00.339825 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" event={"ID":"6b725a2e-30e9-455e-b359-9f0ac33eaf00","Type":"ContainerDied","Data":"7bd33a3aaa34f511ff2a890b3ed6ce4955f5f3e3c1d65214160e7fab39cbbb71"} Nov 27 11:32:00 crc kubenswrapper[4796]: I1127 11:32:00.339850 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" event={"ID":"6b725a2e-30e9-455e-b359-9f0ac33eaf00","Type":"ContainerStarted","Data":"50a440ff316e8e73bbd6648006a2dfbd33dade9568a722ff33819693e2776693"} Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.695372 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.763073 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b725a2e-30e9-455e-b359-9f0ac33eaf00-config-volume\") pod \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.763241 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qpns\" (UniqueName: \"kubernetes.io/projected/6b725a2e-30e9-455e-b359-9f0ac33eaf00-kube-api-access-9qpns\") pod \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.763277 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b725a2e-30e9-455e-b359-9f0ac33eaf00-secret-volume\") pod \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\" (UID: \"6b725a2e-30e9-455e-b359-9f0ac33eaf00\") " Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.764015 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b725a2e-30e9-455e-b359-9f0ac33eaf00-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b725a2e-30e9-455e-b359-9f0ac33eaf00" (UID: "6b725a2e-30e9-455e-b359-9f0ac33eaf00"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.769397 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b725a2e-30e9-455e-b359-9f0ac33eaf00-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b725a2e-30e9-455e-b359-9f0ac33eaf00" (UID: "6b725a2e-30e9-455e-b359-9f0ac33eaf00"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.769416 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b725a2e-30e9-455e-b359-9f0ac33eaf00-kube-api-access-9qpns" (OuterVolumeSpecName: "kube-api-access-9qpns") pod "6b725a2e-30e9-455e-b359-9f0ac33eaf00" (UID: "6b725a2e-30e9-455e-b359-9f0ac33eaf00"). InnerVolumeSpecName "kube-api-access-9qpns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.864998 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qpns\" (UniqueName: \"kubernetes.io/projected/6b725a2e-30e9-455e-b359-9f0ac33eaf00-kube-api-access-9qpns\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.865038 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b725a2e-30e9-455e-b359-9f0ac33eaf00-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.865048 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b725a2e-30e9-455e-b359-9f0ac33eaf00-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.879220 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:32:01 crc kubenswrapper[4796]: I1127 11:32:01.879373 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:32:02 crc kubenswrapper[4796]: I1127 11:32:02.354128 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" event={"ID":"6b725a2e-30e9-455e-b359-9f0ac33eaf00","Type":"ContainerDied","Data":"50a440ff316e8e73bbd6648006a2dfbd33dade9568a722ff33819693e2776693"} Nov 27 11:32:02 crc kubenswrapper[4796]: I1127 11:32:02.354170 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50a440ff316e8e73bbd6648006a2dfbd33dade9568a722ff33819693e2776693" Nov 27 11:32:02 crc kubenswrapper[4796]: I1127 11:32:02.354196 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj" Nov 27 11:32:02 crc kubenswrapper[4796]: I1127 11:32:02.547297 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 11:32:02 crc kubenswrapper[4796]: I1127 11:32:02.572995 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 11:32:03 crc kubenswrapper[4796]: I1127 11:32:03.110503 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 11:32:03 crc kubenswrapper[4796]: I1127 11:32:03.837624 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 11:32:04 crc kubenswrapper[4796]: I1127 11:32:04.110656 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 11:32:04 crc kubenswrapper[4796]: I1127 11:32:04.120562 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 11:32:04 crc kubenswrapper[4796]: I1127 11:32:04.158674 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 11:32:04 crc kubenswrapper[4796]: I1127 11:32:04.161827 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 11:32:04 crc kubenswrapper[4796]: I1127 11:32:04.630668 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.354513 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pmhfw"] Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.354789 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" podUID="9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" containerName="controller-manager" containerID="cri-o://b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34" gracePeriod=30 Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.477510 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc"] Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.478207 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerName="route-controller-manager" containerID="cri-o://2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927" gracePeriod=30 Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.660444 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.733638 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.820084 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-872x8\" (UniqueName: \"kubernetes.io/projected/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-kube-api-access-872x8\") pod \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.820172 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-config\") pod \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.820217 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-serving-cert\") pod \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.820311 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-proxy-ca-bundles\") pod \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.820347 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-client-ca\") pod \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\" (UID: \"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.821683 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-client-ca" (OuterVolumeSpecName: "client-ca") pod "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" (UID: "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.821835 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" (UID: "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.823187 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-config" (OuterVolumeSpecName: "config") pod "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" (UID: "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.831135 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" (UID: "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.833589 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-kube-api-access-872x8" (OuterVolumeSpecName: "kube-api-access-872x8") pod "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" (UID: "9d7c25bb-ecfa-46d8-acab-c14c2a292bf6"). InnerVolumeSpecName "kube-api-access-872x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.838576 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.841595 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-857459b668-qnn27"] Nov 27 11:32:05 crc kubenswrapper[4796]: E1127 11:32:05.841965 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b725a2e-30e9-455e-b359-9f0ac33eaf00" containerName="collect-profiles" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.841996 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b725a2e-30e9-455e-b359-9f0ac33eaf00" containerName="collect-profiles" Nov 27 11:32:05 crc kubenswrapper[4796]: E1127 11:32:05.842019 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842029 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 11:32:05 crc kubenswrapper[4796]: E1127 11:32:05.842047 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" containerName="controller-manager" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842056 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" containerName="controller-manager" Nov 27 11:32:05 crc kubenswrapper[4796]: E1127 11:32:05.842072 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerName="route-controller-manager" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842083 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerName="route-controller-manager" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842208 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerName="route-controller-manager" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842222 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842238 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b725a2e-30e9-455e-b359-9f0ac33eaf00" containerName="collect-profiles" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.842249 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" containerName="controller-manager" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.843741 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.850816 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-857459b668-qnn27"] Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.921702 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2mbh\" (UniqueName: \"kubernetes.io/projected/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-kube-api-access-l2mbh\") pod \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.921776 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-client-ca\") pod \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.921823 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-serving-cert\") pod \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.921868 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-config\") pod \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\" (UID: \"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6\") " Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922132 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn4jv\" (UniqueName: \"kubernetes.io/projected/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-kube-api-access-fn4jv\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922174 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-proxy-ca-bundles\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922209 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-config\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922322 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-client-ca\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922360 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-serving-cert\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922404 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-872x8\" (UniqueName: \"kubernetes.io/projected/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-kube-api-access-872x8\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922420 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922437 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922453 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922465 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922728 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-client-ca" (OuterVolumeSpecName: "client-ca") pod "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" (UID: "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.922895 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-config" (OuterVolumeSpecName: "config") pod "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" (UID: "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.925276 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-kube-api-access-l2mbh" (OuterVolumeSpecName: "kube-api-access-l2mbh") pod "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" (UID: "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6"). InnerVolumeSpecName "kube-api-access-l2mbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:05 crc kubenswrapper[4796]: I1127 11:32:05.925435 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" (UID: "10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.023872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-serving-cert\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.023958 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn4jv\" (UniqueName: \"kubernetes.io/projected/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-kube-api-access-fn4jv\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.023993 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-proxy-ca-bundles\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.024025 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-config\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.024068 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-client-ca\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.024112 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2mbh\" (UniqueName: \"kubernetes.io/projected/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-kube-api-access-l2mbh\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.024126 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.024138 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.024151 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.025301 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-client-ca\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.025804 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-proxy-ca-bundles\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.026028 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-config\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.029695 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-serving-cert\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.049863 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn4jv\" (UniqueName: \"kubernetes.io/projected/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-kube-api-access-fn4jv\") pod \"controller-manager-857459b668-qnn27\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.169695 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.355746 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-857459b668-qnn27"] Nov 27 11:32:06 crc kubenswrapper[4796]: W1127 11:32:06.364409 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4acbd996_f1d0_4e9f_8ed4_ca9647903b8d.slice/crio-40799b53b16f3820b1d62c06dc10a12efe3785f1a31618fb42ddd278aae7d92e WatchSource:0}: Error finding container 40799b53b16f3820b1d62c06dc10a12efe3785f1a31618fb42ddd278aae7d92e: Status 404 returned error can't find the container with id 40799b53b16f3820b1d62c06dc10a12efe3785f1a31618fb42ddd278aae7d92e Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.379028 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" event={"ID":"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d","Type":"ContainerStarted","Data":"40799b53b16f3820b1d62c06dc10a12efe3785f1a31618fb42ddd278aae7d92e"} Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.380590 4796 generic.go:334] "Generic (PLEG): container finished" podID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" containerID="2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927" exitCode=0 Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.380644 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" event={"ID":"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6","Type":"ContainerDied","Data":"2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927"} Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.380661 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" event={"ID":"10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6","Type":"ContainerDied","Data":"971c959eb07077f071dd9ba2f1923d91fff67607e1ef619a255bba9f522de24a"} Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.380666 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.380684 4796 scope.go:117] "RemoveContainer" containerID="2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.382980 4796 generic.go:334] "Generic (PLEG): container finished" podID="9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" containerID="b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34" exitCode=0 Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.383022 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" event={"ID":"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6","Type":"ContainerDied","Data":"b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34"} Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.383054 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" event={"ID":"9d7c25bb-ecfa-46d8-acab-c14c2a292bf6","Type":"ContainerDied","Data":"9df4e80a5deb85d98ac5855c841723b8731fc275c633931826c85a8444f93d9c"} Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.383055 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pmhfw" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.415499 4796 scope.go:117] "RemoveContainer" containerID="2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927" Nov 27 11:32:06 crc kubenswrapper[4796]: E1127 11:32:06.416690 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927\": container with ID starting with 2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927 not found: ID does not exist" containerID="2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.416724 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927"} err="failed to get container status \"2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927\": rpc error: code = NotFound desc = could not find container \"2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927\": container with ID starting with 2d728c87e303b954fdc082b84f58afde23292c2875cd0217e3b16bae3a50d927 not found: ID does not exist" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.416746 4796 scope.go:117] "RemoveContainer" containerID="b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.417874 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc"] Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.425769 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l8bvc"] Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.434025 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pmhfw"] Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.435348 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pmhfw"] Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.439867 4796 scope.go:117] "RemoveContainer" containerID="b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34" Nov 27 11:32:06 crc kubenswrapper[4796]: E1127 11:32:06.441953 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34\": container with ID starting with b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34 not found: ID does not exist" containerID="b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34" Nov 27 11:32:06 crc kubenswrapper[4796]: I1127 11:32:06.441992 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34"} err="failed to get container status \"b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34\": rpc error: code = NotFound desc = could not find container \"b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34\": container with ID starting with b01bd3cdcc0f48527fb2d64aac2b5086c6991a7d05a3c3f3913f002ca3ca9a34 not found: ID does not exist" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.172376 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.221886 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t"] Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.223113 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.226017 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.226255 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.226450 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.226520 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.226588 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.227579 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.228026 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t"] Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.337999 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-config\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.338046 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rbgf\" (UniqueName: \"kubernetes.io/projected/40af7ef9-40e5-4617-afae-ec3e0fdba647-kube-api-access-6rbgf\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.338093 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40af7ef9-40e5-4617-afae-ec3e0fdba647-serving-cert\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.338176 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-client-ca\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.391127 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" event={"ID":"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d","Type":"ContainerStarted","Data":"ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7"} Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.391327 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.397705 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.414103 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" podStartSLOduration=2.414081885 podStartE2EDuration="2.414081885s" podCreationTimestamp="2025-11-27 11:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:32:07.410813385 +0000 UTC m=+444.929132373" watchObservedRunningTime="2025-11-27 11:32:07.414081885 +0000 UTC m=+444.932400813" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.439948 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-config\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.440230 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rbgf\" (UniqueName: \"kubernetes.io/projected/40af7ef9-40e5-4617-afae-ec3e0fdba647-kube-api-access-6rbgf\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.440366 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40af7ef9-40e5-4617-afae-ec3e0fdba647-serving-cert\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.440505 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-client-ca\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.442024 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-config\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.442240 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-client-ca\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.449166 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40af7ef9-40e5-4617-afae-ec3e0fdba647-serving-cert\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.464573 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rbgf\" (UniqueName: \"kubernetes.io/projected/40af7ef9-40e5-4617-afae-ec3e0fdba647-kube-api-access-6rbgf\") pod \"route-controller-manager-588756b8c7-psj9t\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.536440 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.583734 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6" path="/var/lib/kubelet/pods/10b152a1-4ca1-431b-8dc6-a0f5e1a8c9b6/volumes" Nov 27 11:32:07 crc kubenswrapper[4796]: I1127 11:32:07.584464 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d7c25bb-ecfa-46d8-acab-c14c2a292bf6" path="/var/lib/kubelet/pods/9d7c25bb-ecfa-46d8-acab-c14c2a292bf6/volumes" Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.029650 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t"] Nov 27 11:32:08 crc kubenswrapper[4796]: W1127 11:32:08.032704 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40af7ef9_40e5_4617_afae_ec3e0fdba647.slice/crio-43f7618486f90d44d4266ec538a89dd1538c3d6c5bb74372016a386d328a2637 WatchSource:0}: Error finding container 43f7618486f90d44d4266ec538a89dd1538c3d6c5bb74372016a386d328a2637: Status 404 returned error can't find the container with id 43f7618486f90d44d4266ec538a89dd1538c3d6c5bb74372016a386d328a2637 Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.402585 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" event={"ID":"40af7ef9-40e5-4617-afae-ec3e0fdba647","Type":"ContainerStarted","Data":"32aeb2298bf8d6f7edbf6884035b5429b6cae89f37b036afaa27b4cf368e4990"} Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.404164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" event={"ID":"40af7ef9-40e5-4617-afae-ec3e0fdba647","Type":"ContainerStarted","Data":"43f7618486f90d44d4266ec538a89dd1538c3d6c5bb74372016a386d328a2637"} Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.404338 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.423202 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" podStartSLOduration=3.423185891 podStartE2EDuration="3.423185891s" podCreationTimestamp="2025-11-27 11:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:32:08.420944989 +0000 UTC m=+445.939263917" watchObservedRunningTime="2025-11-27 11:32:08.423185891 +0000 UTC m=+445.941504819" Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.516019 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 11:32:08 crc kubenswrapper[4796]: I1127 11:32:08.779355 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:10 crc kubenswrapper[4796]: I1127 11:32:10.561027 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrwn6"] Nov 27 11:32:10 crc kubenswrapper[4796]: I1127 11:32:10.561858 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hrwn6" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="registry-server" containerID="cri-o://ad93c7a7ffa992ef8e13a58a0921b45dde896687c1fb80467250d073206d70c5" gracePeriod=2 Nov 27 11:32:10 crc kubenswrapper[4796]: I1127 11:32:10.727417 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.421015 4796 generic.go:334] "Generic (PLEG): container finished" podID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerID="ad93c7a7ffa992ef8e13a58a0921b45dde896687c1fb80467250d073206d70c5" exitCode=0 Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.421057 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerDied","Data":"ad93c7a7ffa992ef8e13a58a0921b45dde896687c1fb80467250d073206d70c5"} Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.471531 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.598387 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-catalog-content\") pod \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.598496 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b88q5\" (UniqueName: \"kubernetes.io/projected/244e65ef-52c0-49c2-b1dc-1c6e063d075b-kube-api-access-b88q5\") pod \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.598572 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-utilities\") pod \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\" (UID: \"244e65ef-52c0-49c2-b1dc-1c6e063d075b\") " Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.599874 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-utilities" (OuterVolumeSpecName: "utilities") pod "244e65ef-52c0-49c2-b1dc-1c6e063d075b" (UID: "244e65ef-52c0-49c2-b1dc-1c6e063d075b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.605013 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244e65ef-52c0-49c2-b1dc-1c6e063d075b-kube-api-access-b88q5" (OuterVolumeSpecName: "kube-api-access-b88q5") pod "244e65ef-52c0-49c2-b1dc-1c6e063d075b" (UID: "244e65ef-52c0-49c2-b1dc-1c6e063d075b"). InnerVolumeSpecName "kube-api-access-b88q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.645620 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "244e65ef-52c0-49c2-b1dc-1c6e063d075b" (UID: "244e65ef-52c0-49c2-b1dc-1c6e063d075b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.699863 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.699998 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244e65ef-52c0-49c2-b1dc-1c6e063d075b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:11 crc kubenswrapper[4796]: I1127 11:32:11.700029 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b88q5\" (UniqueName: \"kubernetes.io/projected/244e65ef-52c0-49c2-b1dc-1c6e063d075b-kube-api-access-b88q5\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.160850 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-djbsd"] Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.161125 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-djbsd" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="registry-server" containerID="cri-o://b3709bc2f8a23ebe4c9a8af046625474a487a3e20ed34df4d9975cfd4434740a" gracePeriod=2 Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.431085 4796 generic.go:334] "Generic (PLEG): container finished" podID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerID="b3709bc2f8a23ebe4c9a8af046625474a487a3e20ed34df4d9975cfd4434740a" exitCode=0 Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.431147 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerDied","Data":"b3709bc2f8a23ebe4c9a8af046625474a487a3e20ed34df4d9975cfd4434740a"} Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.436549 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrwn6" event={"ID":"244e65ef-52c0-49c2-b1dc-1c6e063d075b","Type":"ContainerDied","Data":"118cb7834bb7b5fd79f924415132c49b510916dc7c63dda27568f45eb6a4fdbc"} Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.436591 4796 scope.go:117] "RemoveContainer" containerID="ad93c7a7ffa992ef8e13a58a0921b45dde896687c1fb80467250d073206d70c5" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.436725 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrwn6" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.478114 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrwn6"] Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.481129 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hrwn6"] Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.482830 4796 scope.go:117] "RemoveContainer" containerID="dde1bf859ced632eff9237f78321337b0adea5db7e88917249c1a008daab9e7e" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.509253 4796 scope.go:117] "RemoveContainer" containerID="42c764fc7848cd120b2f5dfd696dcade5e79e264e543f28cf75c02ac23ffc25c" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.624454 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.711791 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-catalog-content\") pod \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.711913 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljsm2\" (UniqueName: \"kubernetes.io/projected/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-kube-api-access-ljsm2\") pod \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.711943 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-utilities\") pod \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\" (UID: \"59533ee5-3cf6-43d4-a75e-9611c9c84cb2\") " Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.712719 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-utilities" (OuterVolumeSpecName: "utilities") pod "59533ee5-3cf6-43d4-a75e-9611c9c84cb2" (UID: "59533ee5-3cf6-43d4-a75e-9611c9c84cb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.716001 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-kube-api-access-ljsm2" (OuterVolumeSpecName: "kube-api-access-ljsm2") pod "59533ee5-3cf6-43d4-a75e-9611c9c84cb2" (UID: "59533ee5-3cf6-43d4-a75e-9611c9c84cb2"). InnerVolumeSpecName "kube-api-access-ljsm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.737839 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59533ee5-3cf6-43d4-a75e-9611c9c84cb2" (UID: "59533ee5-3cf6-43d4-a75e-9611c9c84cb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.813802 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.813854 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljsm2\" (UniqueName: \"kubernetes.io/projected/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-kube-api-access-ljsm2\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:12 crc kubenswrapper[4796]: I1127 11:32:12.813879 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59533ee5-3cf6-43d4-a75e-9611c9c84cb2-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.163716 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2j2sz"] Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.164011 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2j2sz" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="registry-server" containerID="cri-o://5bbd3707748056bdc1110499fbfd78cfdabfd6e0ed331fc2a5a742d0a2b88b1b" gracePeriod=2 Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.445420 4796 generic.go:334] "Generic (PLEG): container finished" podID="6ee15471-b388-4e39-adca-a7409e0e90be" containerID="5bbd3707748056bdc1110499fbfd78cfdabfd6e0ed331fc2a5a742d0a2b88b1b" exitCode=0 Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.445455 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerDied","Data":"5bbd3707748056bdc1110499fbfd78cfdabfd6e0ed331fc2a5a742d0a2b88b1b"} Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.450944 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-djbsd" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.451529 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-djbsd" event={"ID":"59533ee5-3cf6-43d4-a75e-9611c9c84cb2","Type":"ContainerDied","Data":"39f47c0a521e0a2b51a2f25a3e36b1a204c44f29476a543dbd0358d4b024aef2"} Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.451606 4796 scope.go:117] "RemoveContainer" containerID="b3709bc2f8a23ebe4c9a8af046625474a487a3e20ed34df4d9975cfd4434740a" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.467773 4796 scope.go:117] "RemoveContainer" containerID="78b69ab0ef2d4ffb9a2e80e800e71a8ed1171353975e169b9ff5084834637056" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.503473 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-djbsd"] Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.507032 4796 scope.go:117] "RemoveContainer" containerID="1ebab4aaf456e5d5b3195cf733547d4ad92691175f6c838daad48b884a7d4c8e" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.507162 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-djbsd"] Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.579011 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" path="/var/lib/kubelet/pods/244e65ef-52c0-49c2-b1dc-1c6e063d075b/volumes" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.580107 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" path="/var/lib/kubelet/pods/59533ee5-3cf6-43d4-a75e-9611c9c84cb2/volumes" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.626281 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.724817 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-utilities\") pod \"6ee15471-b388-4e39-adca-a7409e0e90be\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.724865 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-catalog-content\") pod \"6ee15471-b388-4e39-adca-a7409e0e90be\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.724937 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsbmn\" (UniqueName: \"kubernetes.io/projected/6ee15471-b388-4e39-adca-a7409e0e90be-kube-api-access-gsbmn\") pod \"6ee15471-b388-4e39-adca-a7409e0e90be\" (UID: \"6ee15471-b388-4e39-adca-a7409e0e90be\") " Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.727153 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-utilities" (OuterVolumeSpecName: "utilities") pod "6ee15471-b388-4e39-adca-a7409e0e90be" (UID: "6ee15471-b388-4e39-adca-a7409e0e90be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.775744 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee15471-b388-4e39-adca-a7409e0e90be-kube-api-access-gsbmn" (OuterVolumeSpecName: "kube-api-access-gsbmn") pod "6ee15471-b388-4e39-adca-a7409e0e90be" (UID: "6ee15471-b388-4e39-adca-a7409e0e90be"). InnerVolumeSpecName "kube-api-access-gsbmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.816387 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ee15471-b388-4e39-adca-a7409e0e90be" (UID: "6ee15471-b388-4e39-adca-a7409e0e90be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.826116 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsbmn\" (UniqueName: \"kubernetes.io/projected/6ee15471-b388-4e39-adca-a7409e0e90be-kube-api-access-gsbmn\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.826146 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:13 crc kubenswrapper[4796]: I1127 11:32:13.826158 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ee15471-b388-4e39-adca-a7409e0e90be-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.460376 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j2sz" event={"ID":"6ee15471-b388-4e39-adca-a7409e0e90be","Type":"ContainerDied","Data":"e512b060ec2304d67578336b3b024e4ee12c873751b513649931da5beccf05c5"} Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.460416 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j2sz" Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.460443 4796 scope.go:117] "RemoveContainer" containerID="5bbd3707748056bdc1110499fbfd78cfdabfd6e0ed331fc2a5a742d0a2b88b1b" Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.477635 4796 scope.go:117] "RemoveContainer" containerID="a1b81c6a6daa4769044a0ddaa6ed3b628e346b4e7d29adc1d5ba32870b43e902" Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.494595 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2j2sz"] Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.497953 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2j2sz"] Nov 27 11:32:14 crc kubenswrapper[4796]: I1127 11:32:14.498722 4796 scope.go:117] "RemoveContainer" containerID="9c35dd07466aadf358ceedefd562d8a4374afe88c3b3a3a4f706257a30594b22" Nov 27 11:32:15 crc kubenswrapper[4796]: I1127 11:32:15.580417 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" path="/var/lib/kubelet/pods/6ee15471-b388-4e39-adca-a7409e0e90be/volumes" Nov 27 11:32:31 crc kubenswrapper[4796]: I1127 11:32:31.879549 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:32:31 crc kubenswrapper[4796]: I1127 11:32:31.880318 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:32:31 crc kubenswrapper[4796]: I1127 11:32:31.880385 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:32:31 crc kubenswrapper[4796]: I1127 11:32:31.881239 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2bc2f5728589faed16620248b20e0a440b5a67ed778b4753d5c8d2dca98cb558"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:32:31 crc kubenswrapper[4796]: I1127 11:32:31.881359 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://2bc2f5728589faed16620248b20e0a440b5a67ed778b4753d5c8d2dca98cb558" gracePeriod=600 Nov 27 11:32:32 crc kubenswrapper[4796]: I1127 11:32:32.585807 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="2bc2f5728589faed16620248b20e0a440b5a67ed778b4753d5c8d2dca98cb558" exitCode=0 Nov 27 11:32:32 crc kubenswrapper[4796]: I1127 11:32:32.585947 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"2bc2f5728589faed16620248b20e0a440b5a67ed778b4753d5c8d2dca98cb558"} Nov 27 11:32:32 crc kubenswrapper[4796]: I1127 11:32:32.586333 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"39b36d5b8e5bafc99dea0a6943dc340119f3d41fbf1133a259918c397829bd29"} Nov 27 11:32:32 crc kubenswrapper[4796]: I1127 11:32:32.586372 4796 scope.go:117] "RemoveContainer" containerID="ce5aefba2e68283b86ffacd6777bf705e0552665e9bc5a75a22edd780c3d122e" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.375704 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t"] Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.376663 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" podUID="40af7ef9-40e5-4617-afae-ec3e0fdba647" containerName="route-controller-manager" containerID="cri-o://32aeb2298bf8d6f7edbf6884035b5429b6cae89f37b036afaa27b4cf368e4990" gracePeriod=30 Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.667862 4796 generic.go:334] "Generic (PLEG): container finished" podID="40af7ef9-40e5-4617-afae-ec3e0fdba647" containerID="32aeb2298bf8d6f7edbf6884035b5429b6cae89f37b036afaa27b4cf368e4990" exitCode=0 Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.667984 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" event={"ID":"40af7ef9-40e5-4617-afae-ec3e0fdba647","Type":"ContainerDied","Data":"32aeb2298bf8d6f7edbf6884035b5429b6cae89f37b036afaa27b4cf368e4990"} Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.764467 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.856308 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40af7ef9-40e5-4617-afae-ec3e0fdba647-serving-cert\") pod \"40af7ef9-40e5-4617-afae-ec3e0fdba647\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.856379 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rbgf\" (UniqueName: \"kubernetes.io/projected/40af7ef9-40e5-4617-afae-ec3e0fdba647-kube-api-access-6rbgf\") pod \"40af7ef9-40e5-4617-afae-ec3e0fdba647\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.856408 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-client-ca\") pod \"40af7ef9-40e5-4617-afae-ec3e0fdba647\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.856432 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-config\") pod \"40af7ef9-40e5-4617-afae-ec3e0fdba647\" (UID: \"40af7ef9-40e5-4617-afae-ec3e0fdba647\") " Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.857764 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-client-ca" (OuterVolumeSpecName: "client-ca") pod "40af7ef9-40e5-4617-afae-ec3e0fdba647" (UID: "40af7ef9-40e5-4617-afae-ec3e0fdba647"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.858421 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-config" (OuterVolumeSpecName: "config") pod "40af7ef9-40e5-4617-afae-ec3e0fdba647" (UID: "40af7ef9-40e5-4617-afae-ec3e0fdba647"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.863363 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40af7ef9-40e5-4617-afae-ec3e0fdba647-kube-api-access-6rbgf" (OuterVolumeSpecName: "kube-api-access-6rbgf") pod "40af7ef9-40e5-4617-afae-ec3e0fdba647" (UID: "40af7ef9-40e5-4617-afae-ec3e0fdba647"). InnerVolumeSpecName "kube-api-access-6rbgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.863499 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40af7ef9-40e5-4617-afae-ec3e0fdba647-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "40af7ef9-40e5-4617-afae-ec3e0fdba647" (UID: "40af7ef9-40e5-4617-afae-ec3e0fdba647"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.957701 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40af7ef9-40e5-4617-afae-ec3e0fdba647-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.957732 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rbgf\" (UniqueName: \"kubernetes.io/projected/40af7ef9-40e5-4617-afae-ec3e0fdba647-kube-api-access-6rbgf\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.957741 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:45 crc kubenswrapper[4796]: I1127 11:32:45.957750 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40af7ef9-40e5-4617-afae-ec3e0fdba647-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:32:46 crc kubenswrapper[4796]: I1127 11:32:46.675581 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" event={"ID":"40af7ef9-40e5-4617-afae-ec3e0fdba647","Type":"ContainerDied","Data":"43f7618486f90d44d4266ec538a89dd1538c3d6c5bb74372016a386d328a2637"} Nov 27 11:32:46 crc kubenswrapper[4796]: I1127 11:32:46.675655 4796 scope.go:117] "RemoveContainer" containerID="32aeb2298bf8d6f7edbf6884035b5429b6cae89f37b036afaa27b4cf368e4990" Nov 27 11:32:46 crc kubenswrapper[4796]: I1127 11:32:46.675684 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t" Nov 27 11:32:46 crc kubenswrapper[4796]: I1127 11:32:46.718593 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t"] Nov 27 11:32:46 crc kubenswrapper[4796]: I1127 11:32:46.719317 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-588756b8c7-psj9t"] Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245432 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976644968-mpj58"] Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245835 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40af7ef9-40e5-4617-afae-ec3e0fdba647" containerName="route-controller-manager" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245846 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="40af7ef9-40e5-4617-afae-ec3e0fdba647" containerName="route-controller-manager" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245857 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="extract-content" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245863 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="extract-content" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245871 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="extract-utilities" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245877 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="extract-utilities" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245886 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245892 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245902 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="extract-utilities" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245907 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="extract-utilities" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245916 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="extract-content" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245921 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="extract-content" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245931 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245937 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245944 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245949 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245956 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="extract-utilities" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245961 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="extract-utilities" Nov 27 11:32:47 crc kubenswrapper[4796]: E1127 11:32:47.245970 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="extract-content" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.245976 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="extract-content" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.246054 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="244e65ef-52c0-49c2-b1dc-1c6e063d075b" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.246064 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="59533ee5-3cf6-43d4-a75e-9611c9c84cb2" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.246074 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="40af7ef9-40e5-4617-afae-ec3e0fdba647" containerName="route-controller-manager" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.246085 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee15471-b388-4e39-adca-a7409e0e90be" containerName="registry-server" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.246435 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.248677 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.249784 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.249858 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.249944 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.250083 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.250741 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.258375 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976644968-mpj58"] Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.274155 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c054fc1-68ae-43a3-9a50-98e58e5b0936-config\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.274196 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5mr5\" (UniqueName: \"kubernetes.io/projected/9c054fc1-68ae-43a3-9a50-98e58e5b0936-kube-api-access-p5mr5\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.274234 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c054fc1-68ae-43a3-9a50-98e58e5b0936-client-ca\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.274254 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c054fc1-68ae-43a3-9a50-98e58e5b0936-serving-cert\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.375699 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5mr5\" (UniqueName: \"kubernetes.io/projected/9c054fc1-68ae-43a3-9a50-98e58e5b0936-kube-api-access-p5mr5\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.375975 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c054fc1-68ae-43a3-9a50-98e58e5b0936-client-ca\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.376065 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c054fc1-68ae-43a3-9a50-98e58e5b0936-serving-cert\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.376373 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c054fc1-68ae-43a3-9a50-98e58e5b0936-config\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.377771 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c054fc1-68ae-43a3-9a50-98e58e5b0936-client-ca\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.378515 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c054fc1-68ae-43a3-9a50-98e58e5b0936-config\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.385845 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c054fc1-68ae-43a3-9a50-98e58e5b0936-serving-cert\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.397992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5mr5\" (UniqueName: \"kubernetes.io/projected/9c054fc1-68ae-43a3-9a50-98e58e5b0936-kube-api-access-p5mr5\") pod \"route-controller-manager-7976644968-mpj58\" (UID: \"9c054fc1-68ae-43a3-9a50-98e58e5b0936\") " pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.563610 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:47 crc kubenswrapper[4796]: I1127 11:32:47.580046 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40af7ef9-40e5-4617-afae-ec3e0fdba647" path="/var/lib/kubelet/pods/40af7ef9-40e5-4617-afae-ec3e0fdba647/volumes" Nov 27 11:32:48 crc kubenswrapper[4796]: I1127 11:32:48.053812 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976644968-mpj58"] Nov 27 11:32:48 crc kubenswrapper[4796]: W1127 11:32:48.058326 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c054fc1_68ae_43a3_9a50_98e58e5b0936.slice/crio-80b4d857a1a29514c6621a61856c3a67313abd5858f7c83bf64fe33a19af4c19 WatchSource:0}: Error finding container 80b4d857a1a29514c6621a61856c3a67313abd5858f7c83bf64fe33a19af4c19: Status 404 returned error can't find the container with id 80b4d857a1a29514c6621a61856c3a67313abd5858f7c83bf64fe33a19af4c19 Nov 27 11:32:48 crc kubenswrapper[4796]: I1127 11:32:48.699362 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" event={"ID":"9c054fc1-68ae-43a3-9a50-98e58e5b0936","Type":"ContainerStarted","Data":"7a2e6ca81ff40ab95b9b79c4a16ac5208cc4ea4eedbdb10516706a7fdd860fb5"} Nov 27 11:32:48 crc kubenswrapper[4796]: I1127 11:32:48.699728 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" event={"ID":"9c054fc1-68ae-43a3-9a50-98e58e5b0936","Type":"ContainerStarted","Data":"80b4d857a1a29514c6621a61856c3a67313abd5858f7c83bf64fe33a19af4c19"} Nov 27 11:32:48 crc kubenswrapper[4796]: I1127 11:32:48.700727 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:48 crc kubenswrapper[4796]: I1127 11:32:48.706121 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" Nov 27 11:32:48 crc kubenswrapper[4796]: I1127 11:32:48.720418 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7976644968-mpj58" podStartSLOduration=3.720403501 podStartE2EDuration="3.720403501s" podCreationTimestamp="2025-11-27 11:32:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:32:48.717676776 +0000 UTC m=+486.235995694" watchObservedRunningTime="2025-11-27 11:32:48.720403501 +0000 UTC m=+486.238722419" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.351828 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-857459b668-qnn27"] Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.353249 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" podUID="4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" containerName="controller-manager" containerID="cri-o://ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7" gracePeriod=30 Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.731251 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.811838 4796 generic.go:334] "Generic (PLEG): container finished" podID="4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" containerID="ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7" exitCode=0 Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.811890 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" event={"ID":"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d","Type":"ContainerDied","Data":"ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7"} Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.811921 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" event={"ID":"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d","Type":"ContainerDied","Data":"40799b53b16f3820b1d62c06dc10a12efe3785f1a31618fb42ddd278aae7d92e"} Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.811948 4796 scope.go:117] "RemoveContainer" containerID="ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.812064 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-857459b668-qnn27" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.827413 4796 scope.go:117] "RemoveContainer" containerID="ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7" Nov 27 11:33:05 crc kubenswrapper[4796]: E1127 11:33:05.828353 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7\": container with ID starting with ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7 not found: ID does not exist" containerID="ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.828406 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7"} err="failed to get container status \"ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7\": rpc error: code = NotFound desc = could not find container \"ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7\": container with ID starting with ebfa41cc71c748ebbb3d10e17da78b4ffa18ac5562c3e6aa0c53b09da8bc40a7 not found: ID does not exist" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.911888 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-serving-cert\") pod \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.912021 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-proxy-ca-bundles\") pod \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.912095 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-config\") pod \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.912224 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn4jv\" (UniqueName: \"kubernetes.io/projected/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-kube-api-access-fn4jv\") pod \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.912325 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-client-ca\") pod \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\" (UID: \"4acbd996-f1d0-4e9f-8ed4-ca9647903b8d\") " Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.913455 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-client-ca" (OuterVolumeSpecName: "client-ca") pod "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" (UID: "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.913440 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" (UID: "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.913509 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-config" (OuterVolumeSpecName: "config") pod "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" (UID: "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.918750 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-kube-api-access-fn4jv" (OuterVolumeSpecName: "kube-api-access-fn4jv") pod "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" (UID: "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d"). InnerVolumeSpecName "kube-api-access-fn4jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:33:05 crc kubenswrapper[4796]: I1127 11:33:05.919781 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" (UID: "4acbd996-f1d0-4e9f-8ed4-ca9647903b8d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.013981 4796 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.014037 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.014053 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn4jv\" (UniqueName: \"kubernetes.io/projected/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-kube-api-access-fn4jv\") on node \"crc\" DevicePath \"\"" Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.014072 4796 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.014190 4796 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.147829 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-857459b668-qnn27"] Nov 27 11:33:06 crc kubenswrapper[4796]: I1127 11:33:06.154146 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-857459b668-qnn27"] Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.258867 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-576bd996d7-dfd7g"] Nov 27 11:33:07 crc kubenswrapper[4796]: E1127 11:33:07.259371 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" containerName="controller-manager" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.259383 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" containerName="controller-manager" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.259507 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" containerName="controller-manager" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.259890 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.262003 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.262882 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.262954 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.263049 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.263781 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.272939 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.273560 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.275418 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-576bd996d7-dfd7g"] Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.337539 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-proxy-ca-bundles\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.337613 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-serving-cert\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.337750 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6r2m\" (UniqueName: \"kubernetes.io/projected/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-kube-api-access-q6r2m\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.337808 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-config\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.337843 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-client-ca\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.438820 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6r2m\" (UniqueName: \"kubernetes.io/projected/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-kube-api-access-q6r2m\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.438902 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-config\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.438967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-client-ca\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.439003 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-proxy-ca-bundles\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.439039 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-serving-cert\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.440545 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-client-ca\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.440559 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-proxy-ca-bundles\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.441015 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-config\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.455618 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-serving-cert\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.463901 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6r2m\" (UniqueName: \"kubernetes.io/projected/3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1-kube-api-access-q6r2m\") pod \"controller-manager-576bd996d7-dfd7g\" (UID: \"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1\") " pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.582559 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4acbd996-f1d0-4e9f-8ed4-ca9647903b8d" path="/var/lib/kubelet/pods/4acbd996-f1d0-4e9f-8ed4-ca9647903b8d/volumes" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.588427 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.793831 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-576bd996d7-dfd7g"] Nov 27 11:33:07 crc kubenswrapper[4796]: I1127 11:33:07.833066 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" event={"ID":"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1","Type":"ContainerStarted","Data":"c0247c5fc76c4294e27b1d5c225e6c85641b495f21d3d33f6d83c2c50e315c46"} Nov 27 11:33:08 crc kubenswrapper[4796]: I1127 11:33:08.839643 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" event={"ID":"3b7b1cc5-a9e2-4ab4-80af-5cc1479fe0e1","Type":"ContainerStarted","Data":"205605a8b1bfd611c09bd3c55ffce65e326a51ce1d7853e6815b3e5bf1636273"} Nov 27 11:33:08 crc kubenswrapper[4796]: I1127 11:33:08.840095 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:08 crc kubenswrapper[4796]: I1127 11:33:08.845062 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" Nov 27 11:33:08 crc kubenswrapper[4796]: I1127 11:33:08.858346 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-576bd996d7-dfd7g" podStartSLOduration=3.858326967 podStartE2EDuration="3.858326967s" podCreationTimestamp="2025-11-27 11:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:33:08.855463028 +0000 UTC m=+506.373781966" watchObservedRunningTime="2025-11-27 11:33:08.858326967 +0000 UTC m=+506.376645895" Nov 27 11:33:40 crc kubenswrapper[4796]: I1127 11:33:40.116442 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sngfr"] Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.150375 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerName="oauth-openshift" containerID="cri-o://8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5" gracePeriod=15 Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.628662 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.666811 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79f8586974-fqjbk"] Nov 27 11:34:05 crc kubenswrapper[4796]: E1127 11:34:05.667046 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerName="oauth-openshift" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.667061 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerName="oauth-openshift" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.667205 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerName="oauth-openshift" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.667739 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.690293 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79f8586974-fqjbk"] Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738492 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-provider-selection\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738539 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-login\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738562 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-error\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738597 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-ocp-branding-template\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738616 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-dir\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738645 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-idp-0-file-data\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738676 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-session\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738705 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-service-ca\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738726 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-policies\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738750 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-serving-cert\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738765 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-trusted-ca-bundle\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738785 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-cliconfig\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738805 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv6l4\" (UniqueName: \"kubernetes.io/projected/7aa958d0-0f0b-46c2-884f-e7f23431c022-kube-api-access-zv6l4\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.738841 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-router-certs\") pod \"7aa958d0-0f0b-46c2-884f-e7f23431c022\" (UID: \"7aa958d0-0f0b-46c2-884f-e7f23431c022\") " Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.739594 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.739808 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.740359 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.740453 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.740623 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.744611 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.744855 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.745274 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.746495 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.746989 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.747205 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.747619 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.747664 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa958d0-0f0b-46c2-884f-e7f23431c022-kube-api-access-zv6l4" (OuterVolumeSpecName: "kube-api-access-zv6l4") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "kube-api-access-zv6l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.749514 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7aa958d0-0f0b-46c2-884f-e7f23431c022" (UID: "7aa958d0-0f0b-46c2-884f-e7f23431c022"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840023 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dacc63b3-4351-4beb-a413-1cfafb8d4e94-audit-dir\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840090 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840555 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840622 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-session\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840656 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840677 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840696 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840714 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmmz6\" (UniqueName: \"kubernetes.io/projected/dacc63b3-4351-4beb-a413-1cfafb8d4e94-kube-api-access-rmmz6\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840745 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-login\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840818 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840900 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-router-certs\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.840936 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-error\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841016 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-service-ca\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841046 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-audit-policies\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841127 4796 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841140 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841150 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841159 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841207 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv6l4\" (UniqueName: \"kubernetes.io/projected/7aa958d0-0f0b-46c2-884f-e7f23431c022-kube-api-access-zv6l4\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841221 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841230 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841240 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841249 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841258 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841285 4796 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7aa958d0-0f0b-46c2-884f-e7f23431c022-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841295 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841303 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.841314 4796 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7aa958d0-0f0b-46c2-884f-e7f23431c022-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.942395 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.942530 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-session\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.942614 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943343 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943411 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943453 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmmz6\" (UniqueName: \"kubernetes.io/projected/dacc63b3-4351-4beb-a413-1cfafb8d4e94-kube-api-access-rmmz6\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-login\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943529 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943631 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-router-certs\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943672 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-error\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943700 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943735 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-service-ca\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943777 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-audit-policies\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943818 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dacc63b3-4351-4beb-a413-1cfafb8d4e94-audit-dir\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.943862 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.944131 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dacc63b3-4351-4beb-a413-1cfafb8d4e94-audit-dir\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.944741 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-service-ca\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.945331 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-audit-policies\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.945449 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.946389 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.948321 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.948501 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.948616 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-login\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.949160 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.949261 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-user-template-error\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.949348 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-router-certs\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.950919 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dacc63b3-4351-4beb-a413-1cfafb8d4e94-v4-0-config-system-session\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.959886 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmmz6\" (UniqueName: \"kubernetes.io/projected/dacc63b3-4351-4beb-a413-1cfafb8d4e94-kube-api-access-rmmz6\") pod \"oauth-openshift-79f8586974-fqjbk\" (UID: \"dacc63b3-4351-4beb-a413-1cfafb8d4e94\") " pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:05 crc kubenswrapper[4796]: I1127 11:34:05.996022 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.242951 4796 generic.go:334] "Generic (PLEG): container finished" podID="7aa958d0-0f0b-46c2-884f-e7f23431c022" containerID="8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5" exitCode=0 Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.242995 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" event={"ID":"7aa958d0-0f0b-46c2-884f-e7f23431c022","Type":"ContainerDied","Data":"8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5"} Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.243015 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.243031 4796 scope.go:117] "RemoveContainer" containerID="8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5" Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.243021 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sngfr" event={"ID":"7aa958d0-0f0b-46c2-884f-e7f23431c022","Type":"ContainerDied","Data":"9013135d2607b29eb779e7575ab55a97f90ea835d8399a563f52aa1752dc3a33"} Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.279658 4796 scope.go:117] "RemoveContainer" containerID="8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5" Nov 27 11:34:06 crc kubenswrapper[4796]: E1127 11:34:06.282874 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5\": container with ID starting with 8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5 not found: ID does not exist" containerID="8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5" Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.282948 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5"} err="failed to get container status \"8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5\": rpc error: code = NotFound desc = could not find container \"8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5\": container with ID starting with 8563bd0d7bc3a032be150b791c3e4e9043c2ee985f79bb4e7d9daccc63255ff5 not found: ID does not exist" Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.291857 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sngfr"] Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.298634 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sngfr"] Nov 27 11:34:06 crc kubenswrapper[4796]: I1127 11:34:06.405473 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79f8586974-fqjbk"] Nov 27 11:34:07 crc kubenswrapper[4796]: I1127 11:34:07.252365 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" event={"ID":"dacc63b3-4351-4beb-a413-1cfafb8d4e94","Type":"ContainerStarted","Data":"bfeeb8901b29054c4edbd0882b5e33912581a315f69af4e5ff0662594c321ccc"} Nov 27 11:34:07 crc kubenswrapper[4796]: I1127 11:34:07.252796 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" event={"ID":"dacc63b3-4351-4beb-a413-1cfafb8d4e94","Type":"ContainerStarted","Data":"08c6a9d020afee82e956801e8e93389240d0d96f6813532140bdc336c7a0a3f4"} Nov 27 11:34:07 crc kubenswrapper[4796]: I1127 11:34:07.253200 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:07 crc kubenswrapper[4796]: I1127 11:34:07.298178 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" podStartSLOduration=27.298140339 podStartE2EDuration="27.298140339s" podCreationTimestamp="2025-11-27 11:33:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:34:07.288898883 +0000 UTC m=+564.807217851" watchObservedRunningTime="2025-11-27 11:34:07.298140339 +0000 UTC m=+564.816459267" Nov 27 11:34:07 crc kubenswrapper[4796]: I1127 11:34:07.416461 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79f8586974-fqjbk" Nov 27 11:34:07 crc kubenswrapper[4796]: I1127 11:34:07.575630 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aa958d0-0f0b-46c2-884f-e7f23431c022" path="/var/lib/kubelet/pods/7aa958d0-0f0b-46c2-884f-e7f23431c022/volumes" Nov 27 11:35:01 crc kubenswrapper[4796]: I1127 11:35:01.879901 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:35:01 crc kubenswrapper[4796]: I1127 11:35:01.880828 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:35:31 crc kubenswrapper[4796]: I1127 11:35:31.879418 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:35:31 crc kubenswrapper[4796]: I1127 11:35:31.880185 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.818356 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tddgt"] Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.819244 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tddgt" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="registry-server" containerID="cri-o://8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6" gracePeriod=30 Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.824629 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxkmw"] Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.824828 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hxkmw" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="registry-server" containerID="cri-o://c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce" gracePeriod=30 Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.838618 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fdwzj"] Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.839175 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" containerID="cri-o://e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d" gracePeriod=30 Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.846946 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbl4f"] Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.848184 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pbl4f" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="registry-server" containerID="cri-o://34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" gracePeriod=30 Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.857900 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lv29n"] Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.859531 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.863904 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-92gc9"] Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.864166 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-92gc9" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="registry-server" containerID="cri-o://de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4" gracePeriod=30 Nov 27 11:35:38 crc kubenswrapper[4796]: I1127 11:35:38.873091 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lv29n"] Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.038879 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.038938 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.038983 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6crk\" (UniqueName: \"kubernetes.io/projected/097c089f-7300-463f-abc6-2f00735c18d9-kube-api-access-m6crk\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: E1127 11:35:39.047374 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc is running failed: container process not found" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 11:35:39 crc kubenswrapper[4796]: E1127 11:35:39.047890 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc is running failed: container process not found" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 11:35:39 crc kubenswrapper[4796]: E1127 11:35:39.048158 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc is running failed: container process not found" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 11:35:39 crc kubenswrapper[4796]: E1127 11:35:39.048185 4796 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-pbl4f" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="registry-server" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.141147 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.141225 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.142486 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6crk\" (UniqueName: \"kubernetes.io/projected/097c089f-7300-463f-abc6-2f00735c18d9-kube-api-access-m6crk\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.142767 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.151639 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.159446 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6crk\" (UniqueName: \"kubernetes.io/projected/097c089f-7300-463f-abc6-2f00735c18d9-kube-api-access-m6crk\") pod \"marketplace-operator-79b997595-lv29n\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.227365 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.334190 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.339792 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fdwzj_dc437d52-8b8b-430f-bf9e-67492c365aad/marketplace-operator/1.log" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.339852 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.343830 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-trusted-ca\") pod \"dc437d52-8b8b-430f-bf9e-67492c365aad\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.343855 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksnfm\" (UniqueName: \"kubernetes.io/projected/906ed244-7663-473c-8495-fe03bd4aa745-kube-api-access-ksnfm\") pod \"906ed244-7663-473c-8495-fe03bd4aa745\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.343873 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-utilities\") pod \"906ed244-7663-473c-8495-fe03bd4aa745\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.343904 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8z87\" (UniqueName: \"kubernetes.io/projected/dc437d52-8b8b-430f-bf9e-67492c365aad-kube-api-access-k8z87\") pod \"dc437d52-8b8b-430f-bf9e-67492c365aad\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.343932 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-catalog-content\") pod \"906ed244-7663-473c-8495-fe03bd4aa745\" (UID: \"906ed244-7663-473c-8495-fe03bd4aa745\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.343969 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-operator-metrics\") pod \"dc437d52-8b8b-430f-bf9e-67492c365aad\" (UID: \"dc437d52-8b8b-430f-bf9e-67492c365aad\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.345036 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "dc437d52-8b8b-430f-bf9e-67492c365aad" (UID: "dc437d52-8b8b-430f-bf9e-67492c365aad"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.346560 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-utilities" (OuterVolumeSpecName: "utilities") pod "906ed244-7663-473c-8495-fe03bd4aa745" (UID: "906ed244-7663-473c-8495-fe03bd4aa745"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.349180 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "dc437d52-8b8b-430f-bf9e-67492c365aad" (UID: "dc437d52-8b8b-430f-bf9e-67492c365aad"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.354527 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/906ed244-7663-473c-8495-fe03bd4aa745-kube-api-access-ksnfm" (OuterVolumeSpecName: "kube-api-access-ksnfm") pod "906ed244-7663-473c-8495-fe03bd4aa745" (UID: "906ed244-7663-473c-8495-fe03bd4aa745"). InnerVolumeSpecName "kube-api-access-ksnfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.362215 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.349593 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc437d52-8b8b-430f-bf9e-67492c365aad-kube-api-access-k8z87" (OuterVolumeSpecName: "kube-api-access-k8z87") pod "dc437d52-8b8b-430f-bf9e-67492c365aad" (UID: "dc437d52-8b8b-430f-bf9e-67492c365aad"). InnerVolumeSpecName "kube-api-access-k8z87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.376787 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.445097 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.445144 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc437d52-8b8b-430f-bf9e-67492c365aad-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.445158 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksnfm\" (UniqueName: \"kubernetes.io/projected/906ed244-7663-473c-8495-fe03bd4aa745-kube-api-access-ksnfm\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.445171 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.445182 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8z87\" (UniqueName: \"kubernetes.io/projected/dc437d52-8b8b-430f-bf9e-67492c365aad-kube-api-access-k8z87\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.472697 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "906ed244-7663-473c-8495-fe03bd4aa745" (UID: "906ed244-7663-473c-8495-fe03bd4aa745"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546101 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-catalog-content\") pod \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546147 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqrfj\" (UniqueName: \"kubernetes.io/projected/08792b59-9118-40df-ab6a-ef002d223f4e-kube-api-access-vqrfj\") pod \"08792b59-9118-40df-ab6a-ef002d223f4e\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546206 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-utilities\") pod \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546232 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr4ph\" (UniqueName: \"kubernetes.io/projected/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-kube-api-access-jr4ph\") pod \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\" (UID: \"83a6d9b2-0bc7-46fb-9528-afce9582aaaf\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546291 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-catalog-content\") pod \"08792b59-9118-40df-ab6a-ef002d223f4e\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546335 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-utilities\") pod \"08792b59-9118-40df-ab6a-ef002d223f4e\" (UID: \"08792b59-9118-40df-ab6a-ef002d223f4e\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.546510 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906ed244-7663-473c-8495-fe03bd4aa745-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.547125 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-utilities" (OuterVolumeSpecName: "utilities") pod "08792b59-9118-40df-ab6a-ef002d223f4e" (UID: "08792b59-9118-40df-ab6a-ef002d223f4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.547766 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-utilities" (OuterVolumeSpecName: "utilities") pod "83a6d9b2-0bc7-46fb-9528-afce9582aaaf" (UID: "83a6d9b2-0bc7-46fb-9528-afce9582aaaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.552939 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08792b59-9118-40df-ab6a-ef002d223f4e-kube-api-access-vqrfj" (OuterVolumeSpecName: "kube-api-access-vqrfj") pod "08792b59-9118-40df-ab6a-ef002d223f4e" (UID: "08792b59-9118-40df-ab6a-ef002d223f4e"). InnerVolumeSpecName "kube-api-access-vqrfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.553005 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-kube-api-access-jr4ph" (OuterVolumeSpecName: "kube-api-access-jr4ph") pod "83a6d9b2-0bc7-46fb-9528-afce9582aaaf" (UID: "83a6d9b2-0bc7-46fb-9528-afce9582aaaf"). InnerVolumeSpecName "kube-api-access-jr4ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.566931 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08792b59-9118-40df-ab6a-ef002d223f4e" (UID: "08792b59-9118-40df-ab6a-ef002d223f4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.610892 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83a6d9b2-0bc7-46fb-9528-afce9582aaaf" (UID: "83a6d9b2-0bc7-46fb-9528-afce9582aaaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.647287 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.647338 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.647351 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqrfj\" (UniqueName: \"kubernetes.io/projected/08792b59-9118-40df-ab6a-ef002d223f4e-kube-api-access-vqrfj\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.647361 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.647372 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr4ph\" (UniqueName: \"kubernetes.io/projected/83a6d9b2-0bc7-46fb-9528-afce9582aaaf-kube-api-access-jr4ph\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.647411 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08792b59-9118-40df-ab6a-ef002d223f4e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.660079 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lv29n"] Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.682327 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.849861 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-utilities\") pod \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.850409 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbbks\" (UniqueName: \"kubernetes.io/projected/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-kube-api-access-nbbks\") pod \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.850526 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-catalog-content\") pod \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\" (UID: \"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d\") " Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.853845 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-utilities" (OuterVolumeSpecName: "utilities") pod "a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" (UID: "a2bcf5d3-efae-40cb-b701-fe0f65a28d0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.858391 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-kube-api-access-nbbks" (OuterVolumeSpecName: "kube-api-access-nbbks") pod "a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" (UID: "a2bcf5d3-efae-40cb-b701-fe0f65a28d0d"). InnerVolumeSpecName "kube-api-access-nbbks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.913524 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" (UID: "a2bcf5d3-efae-40cb-b701-fe0f65a28d0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.931663 4796 generic.go:334] "Generic (PLEG): container finished" podID="08792b59-9118-40df-ab6a-ef002d223f4e" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" exitCode=0 Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.931744 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerDied","Data":"34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.931777 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbl4f" event={"ID":"08792b59-9118-40df-ab6a-ef002d223f4e","Type":"ContainerDied","Data":"365ee2422dd27d8a89ff2ac83a738c141278852cd4bddfd5d7896ee0a282339e"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.931800 4796 scope.go:117] "RemoveContainer" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.931936 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbl4f" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.934673 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" event={"ID":"097c089f-7300-463f-abc6-2f00735c18d9","Type":"ContainerStarted","Data":"c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.934710 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" event={"ID":"097c089f-7300-463f-abc6-2f00735c18d9","Type":"ContainerStarted","Data":"5c851084da1c87d135d804b42531741beb2b37caacfdaaf333d70b7750067679"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.935653 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.938330 4796 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-lv29n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.938398 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" podUID="097c089f-7300-463f-abc6-2f00735c18d9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.938874 4796 generic.go:334] "Generic (PLEG): container finished" podID="906ed244-7663-473c-8495-fe03bd4aa745" containerID="de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4" exitCode=0 Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.939230 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92gc9" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.940384 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerDied","Data":"de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.940553 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92gc9" event={"ID":"906ed244-7663-473c-8495-fe03bd4aa745","Type":"ContainerDied","Data":"3fda918261abca918ce66ce611448f7f1aaeadc87f8ae13249930497eda9fdc7"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.944432 4796 generic.go:334] "Generic (PLEG): container finished" podID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerID="8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6" exitCode=0 Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.944582 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerDied","Data":"8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.944641 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddgt" event={"ID":"a2bcf5d3-efae-40cb-b701-fe0f65a28d0d","Type":"ContainerDied","Data":"f0dcee177b78fc58050674ec6da03d213ccb7a9530f79e2b0e05fc1c60b5cafc"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.945953 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddgt" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.952391 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbbks\" (UniqueName: \"kubernetes.io/projected/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-kube-api-access-nbbks\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.952597 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.952732 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.954750 4796 generic.go:334] "Generic (PLEG): container finished" podID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerID="c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce" exitCode=0 Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.954816 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerDied","Data":"c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.954847 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hxkmw" event={"ID":"83a6d9b2-0bc7-46fb-9528-afce9582aaaf","Type":"ContainerDied","Data":"a0b3651b70bbb09a86187f7e1bf649370fa0476c6063ee38259ee6d2e6b04c95"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.954916 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hxkmw" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.960113 4796 scope.go:117] "RemoveContainer" containerID="b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.960228 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbl4f"] Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.963167 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fdwzj_dc437d52-8b8b-430f-bf9e-67492c365aad/marketplace-operator/1.log" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.963233 4796 generic.go:334] "Generic (PLEG): container finished" podID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerID="e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d" exitCode=0 Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.963342 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerDied","Data":"e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.963382 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" event={"ID":"dc437d52-8b8b-430f-bf9e-67492c365aad","Type":"ContainerDied","Data":"93b39b7b5f804c3efd2862116a665de39994507683dcf1ad9fa62081ddb3da51"} Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.963457 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fdwzj" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.967131 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbl4f"] Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.980415 4796 scope.go:117] "RemoveContainer" containerID="b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4" Nov 27 11:35:39 crc kubenswrapper[4796]: I1127 11:35:39.995176 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" podStartSLOduration=1.995129195 podStartE2EDuration="1.995129195s" podCreationTimestamp="2025-11-27 11:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:35:39.976002757 +0000 UTC m=+657.494321665" watchObservedRunningTime="2025-11-27 11:35:39.995129195 +0000 UTC m=+657.513448123" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.005663 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-92gc9"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.008931 4796 scope.go:117] "RemoveContainer" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.009757 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc\": container with ID starting with 34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc not found: ID does not exist" containerID="34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.009821 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc"} err="failed to get container status \"34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc\": rpc error: code = NotFound desc = could not find container \"34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc\": container with ID starting with 34a55a9d7f8c32ea102f4ebd3f3a2c78ecb1038b98ccde1569d367f4dff313fc not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.009859 4796 scope.go:117] "RemoveContainer" containerID="b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.010617 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa\": container with ID starting with b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa not found: ID does not exist" containerID="b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.010662 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa"} err="failed to get container status \"b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa\": rpc error: code = NotFound desc = could not find container \"b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa\": container with ID starting with b9603736b43e276a846e2af853a0ad76f96c166d3670853e434bebeff26c75fa not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.010687 4796 scope.go:117] "RemoveContainer" containerID="b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.010952 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-92gc9"] Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.012763 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4\": container with ID starting with b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4 not found: ID does not exist" containerID="b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.012812 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4"} err="failed to get container status \"b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4\": rpc error: code = NotFound desc = could not find container \"b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4\": container with ID starting with b304b3f1fbda71ddf1f9aca89178a97251d45ef9d2d3775f1828fe81e06827d4 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.012844 4796 scope.go:117] "RemoveContainer" containerID="de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.017580 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tddgt"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.020699 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tddgt"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.042125 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hxkmw"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.063496 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hxkmw"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.065884 4796 scope.go:117] "RemoveContainer" containerID="65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.070766 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fdwzj"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.073869 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fdwzj"] Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.083699 4796 scope.go:117] "RemoveContainer" containerID="3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.100791 4796 scope.go:117] "RemoveContainer" containerID="de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.101332 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4\": container with ID starting with de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4 not found: ID does not exist" containerID="de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.101372 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4"} err="failed to get container status \"de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4\": rpc error: code = NotFound desc = could not find container \"de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4\": container with ID starting with de6e0ee9eed464c4735b8ff60b73b066e2f6b2400ecbeddb2b69d4bdbb446ea4 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.101400 4796 scope.go:117] "RemoveContainer" containerID="65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.101754 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020\": container with ID starting with 65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020 not found: ID does not exist" containerID="65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.101805 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020"} err="failed to get container status \"65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020\": rpc error: code = NotFound desc = could not find container \"65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020\": container with ID starting with 65bc5999395c8a910ce98f2cc3db3f5d8981ea446ca0347b2ded244617267020 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.101824 4796 scope.go:117] "RemoveContainer" containerID="3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.102099 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d\": container with ID starting with 3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d not found: ID does not exist" containerID="3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.102120 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d"} err="failed to get container status \"3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d\": rpc error: code = NotFound desc = could not find container \"3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d\": container with ID starting with 3bb8ebce1212966673a590f17b93f8ace55f58d408cbb438ec2341b43784b39d not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.102135 4796 scope.go:117] "RemoveContainer" containerID="8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.128218 4796 scope.go:117] "RemoveContainer" containerID="bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.149474 4796 scope.go:117] "RemoveContainer" containerID="a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.162057 4796 scope.go:117] "RemoveContainer" containerID="8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.162423 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6\": container with ID starting with 8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6 not found: ID does not exist" containerID="8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.162458 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6"} err="failed to get container status \"8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6\": rpc error: code = NotFound desc = could not find container \"8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6\": container with ID starting with 8197ee7f75344793a534f278962141207af93d389d50340d2eb85b9fcb7c8da6 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.162484 4796 scope.go:117] "RemoveContainer" containerID="bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.162858 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0\": container with ID starting with bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0 not found: ID does not exist" containerID="bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.162885 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0"} err="failed to get container status \"bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0\": rpc error: code = NotFound desc = could not find container \"bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0\": container with ID starting with bee0dd6544811cc567507c5d216af90de4a343de6010a160bdd40479a8d767a0 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.162902 4796 scope.go:117] "RemoveContainer" containerID="a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.163136 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86\": container with ID starting with a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86 not found: ID does not exist" containerID="a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.163167 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86"} err="failed to get container status \"a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86\": rpc error: code = NotFound desc = could not find container \"a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86\": container with ID starting with a49ace2a484b471fb36a652d554131bd83b7462025289ad1325e5f672a21ee86 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.163185 4796 scope.go:117] "RemoveContainer" containerID="c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.174256 4796 scope.go:117] "RemoveContainer" containerID="4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.191479 4796 scope.go:117] "RemoveContainer" containerID="4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.206839 4796 scope.go:117] "RemoveContainer" containerID="c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.207466 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce\": container with ID starting with c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce not found: ID does not exist" containerID="c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.207528 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce"} err="failed to get container status \"c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce\": rpc error: code = NotFound desc = could not find container \"c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce\": container with ID starting with c8b26276376a06aaa3a66649497b1b9aedf8abc735465f6235ced47e9c6439ce not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.207568 4796 scope.go:117] "RemoveContainer" containerID="4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.207925 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab\": container with ID starting with 4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab not found: ID does not exist" containerID="4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.207954 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab"} err="failed to get container status \"4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab\": rpc error: code = NotFound desc = could not find container \"4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab\": container with ID starting with 4a9edfe4bf2fad0e60b393ec52052a99e002270f318ccbd0658fbcb114dda8ab not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.207972 4796 scope.go:117] "RemoveContainer" containerID="4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.208231 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4\": container with ID starting with 4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4 not found: ID does not exist" containerID="4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.208259 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4"} err="failed to get container status \"4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4\": rpc error: code = NotFound desc = could not find container \"4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4\": container with ID starting with 4f1a08f08f1462a0706f17294e8ada6dc037594d8b880c082143cd9d7e22bac4 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.208293 4796 scope.go:117] "RemoveContainer" containerID="e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.221652 4796 scope.go:117] "RemoveContainer" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.235102 4796 scope.go:117] "RemoveContainer" containerID="e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.235593 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d\": container with ID starting with e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d not found: ID does not exist" containerID="e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.235655 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d"} err="failed to get container status \"e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d\": rpc error: code = NotFound desc = could not find container \"e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d\": container with ID starting with e75a84b14bb681675fda6b16a2052ab7767f1906640af3586063477de78ee20d not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.235687 4796 scope.go:117] "RemoveContainer" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" Nov 27 11:35:40 crc kubenswrapper[4796]: E1127 11:35:40.235994 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420\": container with ID starting with 4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420 not found: ID does not exist" containerID="4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.236014 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420"} err="failed to get container status \"4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420\": rpc error: code = NotFound desc = could not find container \"4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420\": container with ID starting with 4dd2874c2e89da7971d7b1c9a7ff23c5c8ed2de9a02245d7e20aa0f99f9c2420 not found: ID does not exist" Nov 27 11:35:40 crc kubenswrapper[4796]: I1127 11:35:40.982204 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029363 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zjbvh"] Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029541 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029554 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029565 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029571 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029578 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029584 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029592 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029598 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029607 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029613 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029622 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029628 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029640 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029645 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029653 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029658 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029666 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029672 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029680 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029686 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029696 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029702 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029712 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029718 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="extract-utilities" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029726 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029732 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029740 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029746 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: E1127 11:35:41.029756 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029761 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="extract-content" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029841 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029851 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029857 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029863 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029874 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="906ed244-7663-473c-8495-fe03bd4aa745" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029884 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" containerName="registry-server" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.029890 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" containerName="marketplace-operator" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.030530 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.032945 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.042792 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zjbvh"] Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.178066 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt4x7\" (UniqueName: \"kubernetes.io/projected/5e88b470-90dd-4759-939d-65edfde8e6f6-kube-api-access-mt4x7\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.178207 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-catalog-content\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.178334 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-utilities\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.236557 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xpct5"] Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.237981 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.239713 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpct5"] Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.240089 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.279351 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-utilities\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.279395 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt4x7\" (UniqueName: \"kubernetes.io/projected/5e88b470-90dd-4759-939d-65edfde8e6f6-kube-api-access-mt4x7\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.279441 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-catalog-content\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.279766 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-catalog-content\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.279956 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-utilities\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.297730 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt4x7\" (UniqueName: \"kubernetes.io/projected/5e88b470-90dd-4759-939d-65edfde8e6f6-kube-api-access-mt4x7\") pod \"certified-operators-zjbvh\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.357007 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.380544 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5qmc\" (UniqueName: \"kubernetes.io/projected/6b58a0da-df89-4f60-9152-1a56e3604878-kube-api-access-n5qmc\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.380611 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-catalog-content\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.380841 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-utilities\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.482008 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5qmc\" (UniqueName: \"kubernetes.io/projected/6b58a0da-df89-4f60-9152-1a56e3604878-kube-api-access-n5qmc\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.482337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-catalog-content\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.482406 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-utilities\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.482983 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-utilities\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.483831 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-catalog-content\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.507649 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5qmc\" (UniqueName: \"kubernetes.io/projected/6b58a0da-df89-4f60-9152-1a56e3604878-kube-api-access-n5qmc\") pod \"redhat-marketplace-xpct5\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.559183 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.579193 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08792b59-9118-40df-ab6a-ef002d223f4e" path="/var/lib/kubelet/pods/08792b59-9118-40df-ab6a-ef002d223f4e/volumes" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.579955 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a6d9b2-0bc7-46fb-9528-afce9582aaaf" path="/var/lib/kubelet/pods/83a6d9b2-0bc7-46fb-9528-afce9582aaaf/volumes" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.580576 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="906ed244-7663-473c-8495-fe03bd4aa745" path="/var/lib/kubelet/pods/906ed244-7663-473c-8495-fe03bd4aa745/volumes" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.581673 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2bcf5d3-efae-40cb-b701-fe0f65a28d0d" path="/var/lib/kubelet/pods/a2bcf5d3-efae-40cb-b701-fe0f65a28d0d/volumes" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.582467 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc437d52-8b8b-430f-bf9e-67492c365aad" path="/var/lib/kubelet/pods/dc437d52-8b8b-430f-bf9e-67492c365aad/volumes" Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.616133 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zjbvh"] Nov 27 11:35:41 crc kubenswrapper[4796]: W1127 11:35:41.618952 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e88b470_90dd_4759_939d_65edfde8e6f6.slice/crio-c2fd91f9452d3756a5d937cc52377667286165df8ebf030408998c1597ad0420 WatchSource:0}: Error finding container c2fd91f9452d3756a5d937cc52377667286165df8ebf030408998c1597ad0420: Status 404 returned error can't find the container with id c2fd91f9452d3756a5d937cc52377667286165df8ebf030408998c1597ad0420 Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.777703 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpct5"] Nov 27 11:35:41 crc kubenswrapper[4796]: W1127 11:35:41.831487 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b58a0da_df89_4f60_9152_1a56e3604878.slice/crio-6571be8aeff516067da7d9b297bdc9609cc45220ee822093323e59cdeba6b050 WatchSource:0}: Error finding container 6571be8aeff516067da7d9b297bdc9609cc45220ee822093323e59cdeba6b050: Status 404 returned error can't find the container with id 6571be8aeff516067da7d9b297bdc9609cc45220ee822093323e59cdeba6b050 Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.984845 4796 generic.go:334] "Generic (PLEG): container finished" podID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerID="1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b" exitCode=0 Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.984932 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerDied","Data":"1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b"} Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.984974 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerStarted","Data":"c2fd91f9452d3756a5d937cc52377667286165df8ebf030408998c1597ad0420"} Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.988722 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerStarted","Data":"e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf"} Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.988774 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerStarted","Data":"6571be8aeff516067da7d9b297bdc9609cc45220ee822093323e59cdeba6b050"} Nov 27 11:35:41 crc kubenswrapper[4796]: I1127 11:35:41.989789 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:35:42 crc kubenswrapper[4796]: I1127 11:35:42.994299 4796 generic.go:334] "Generic (PLEG): container finished" podID="6b58a0da-df89-4f60-9152-1a56e3604878" containerID="e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf" exitCode=0 Nov 27 11:35:42 crc kubenswrapper[4796]: I1127 11:35:42.994370 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerDied","Data":"e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf"} Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:42.997961 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerStarted","Data":"2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35"} Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.433580 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-66p7v"] Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.435151 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.437208 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.445174 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-66p7v"] Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.608973 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-catalog-content\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.609029 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5qxs\" (UniqueName: \"kubernetes.io/projected/667350f0-abe3-4fc2-8d32-f817daf72697-kube-api-access-n5qxs\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.609535 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-utilities\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.624205 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jxxzb"] Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.625150 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.627122 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.636616 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jxxzb"] Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.710368 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-utilities\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.710439 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5qxs\" (UniqueName: \"kubernetes.io/projected/667350f0-abe3-4fc2-8d32-f817daf72697-kube-api-access-n5qxs\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.710460 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-catalog-content\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.710915 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-catalog-content\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.711007 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-utilities\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.731815 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5qxs\" (UniqueName: \"kubernetes.io/projected/667350f0-abe3-4fc2-8d32-f817daf72697-kube-api-access-n5qxs\") pod \"community-operators-66p7v\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.763536 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.771792 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.812994 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-utilities\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.813475 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-catalog-content\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.813555 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqhv6\" (UniqueName: \"kubernetes.io/projected/79f62242-42c1-459f-9828-00f28d5de84d-kube-api-access-jqhv6\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.914252 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-catalog-content\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.914354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqhv6\" (UniqueName: \"kubernetes.io/projected/79f62242-42c1-459f-9828-00f28d5de84d-kube-api-access-jqhv6\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.914400 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-utilities\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.914774 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-catalog-content\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.914912 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-utilities\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.935353 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqhv6\" (UniqueName: \"kubernetes.io/projected/79f62242-42c1-459f-9828-00f28d5de84d-kube-api-access-jqhv6\") pod \"redhat-operators-jxxzb\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:43 crc kubenswrapper[4796]: I1127 11:35:43.951417 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-66p7v"] Nov 27 11:35:43 crc kubenswrapper[4796]: W1127 11:35:43.974358 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod667350f0_abe3_4fc2_8d32_f817daf72697.slice/crio-6487b9c637ed73666a8a1f6b3004267d73e17c25bca08da5b4decfd180363857 WatchSource:0}: Error finding container 6487b9c637ed73666a8a1f6b3004267d73e17c25bca08da5b4decfd180363857: Status 404 returned error can't find the container with id 6487b9c637ed73666a8a1f6b3004267d73e17c25bca08da5b4decfd180363857 Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.002660 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.004009 4796 generic.go:334] "Generic (PLEG): container finished" podID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerID="2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35" exitCode=0 Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.004051 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerDied","Data":"2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35"} Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.005315 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerStarted","Data":"6487b9c637ed73666a8a1f6b3004267d73e17c25bca08da5b4decfd180363857"} Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.008626 4796 generic.go:334] "Generic (PLEG): container finished" podID="6b58a0da-df89-4f60-9152-1a56e3604878" containerID="6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7" exitCode=0 Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.008669 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerDied","Data":"6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7"} Nov 27 11:35:44 crc kubenswrapper[4796]: I1127 11:35:44.187159 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jxxzb"] Nov 27 11:35:44 crc kubenswrapper[4796]: W1127 11:35:44.223622 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79f62242_42c1_459f_9828_00f28d5de84d.slice/crio-6037a24fbbf813ab07cf808ccfba4524366e67ecb9488778ea08812f82a4e5b4 WatchSource:0}: Error finding container 6037a24fbbf813ab07cf808ccfba4524366e67ecb9488778ea08812f82a4e5b4: Status 404 returned error can't find the container with id 6037a24fbbf813ab07cf808ccfba4524366e67ecb9488778ea08812f82a4e5b4 Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.014609 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerStarted","Data":"8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3"} Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.016444 4796 generic.go:334] "Generic (PLEG): container finished" podID="79f62242-42c1-459f-9828-00f28d5de84d" containerID="c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb" exitCode=0 Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.016509 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerDied","Data":"c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb"} Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.016530 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerStarted","Data":"6037a24fbbf813ab07cf808ccfba4524366e67ecb9488778ea08812f82a4e5b4"} Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.017610 4796 generic.go:334] "Generic (PLEG): container finished" podID="667350f0-abe3-4fc2-8d32-f817daf72697" containerID="c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713" exitCode=0 Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.017636 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerDied","Data":"c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713"} Nov 27 11:35:45 crc kubenswrapper[4796]: I1127 11:35:45.040096 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xpct5" podStartSLOduration=2.244019357 podStartE2EDuration="4.040075916s" podCreationTimestamp="2025-11-27 11:35:41 +0000 UTC" firstStartedPulling="2025-11-27 11:35:42.995768902 +0000 UTC m=+660.514087860" lastFinishedPulling="2025-11-27 11:35:44.791825501 +0000 UTC m=+662.310144419" observedRunningTime="2025-11-27 11:35:45.030542604 +0000 UTC m=+662.548861522" watchObservedRunningTime="2025-11-27 11:35:45.040075916 +0000 UTC m=+662.558394834" Nov 27 11:35:46 crc kubenswrapper[4796]: I1127 11:35:46.023806 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerStarted","Data":"db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f"} Nov 27 11:35:46 crc kubenswrapper[4796]: I1127 11:35:46.025444 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerStarted","Data":"03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a"} Nov 27 11:35:46 crc kubenswrapper[4796]: I1127 11:35:46.027297 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerStarted","Data":"a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21"} Nov 27 11:35:46 crc kubenswrapper[4796]: I1127 11:35:46.074200 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zjbvh" podStartSLOduration=2.074619156 podStartE2EDuration="5.074183803s" podCreationTimestamp="2025-11-27 11:35:41 +0000 UTC" firstStartedPulling="2025-11-27 11:35:41.989583885 +0000 UTC m=+659.507902803" lastFinishedPulling="2025-11-27 11:35:44.989148532 +0000 UTC m=+662.507467450" observedRunningTime="2025-11-27 11:35:46.071854069 +0000 UTC m=+663.590172997" watchObservedRunningTime="2025-11-27 11:35:46.074183803 +0000 UTC m=+663.592502721" Nov 27 11:35:47 crc kubenswrapper[4796]: I1127 11:35:47.035948 4796 generic.go:334] "Generic (PLEG): container finished" podID="79f62242-42c1-459f-9828-00f28d5de84d" containerID="a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21" exitCode=0 Nov 27 11:35:47 crc kubenswrapper[4796]: I1127 11:35:47.036133 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerDied","Data":"a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21"} Nov 27 11:35:47 crc kubenswrapper[4796]: I1127 11:35:47.039480 4796 generic.go:334] "Generic (PLEG): container finished" podID="667350f0-abe3-4fc2-8d32-f817daf72697" containerID="db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f" exitCode=0 Nov 27 11:35:47 crc kubenswrapper[4796]: I1127 11:35:47.039594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerDied","Data":"db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f"} Nov 27 11:35:48 crc kubenswrapper[4796]: I1127 11:35:48.048282 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerStarted","Data":"740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b"} Nov 27 11:35:48 crc kubenswrapper[4796]: I1127 11:35:48.051941 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerStarted","Data":"7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b"} Nov 27 11:35:48 crc kubenswrapper[4796]: I1127 11:35:48.068480 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jxxzb" podStartSLOduration=2.608574053 podStartE2EDuration="5.068463738s" podCreationTimestamp="2025-11-27 11:35:43 +0000 UTC" firstStartedPulling="2025-11-27 11:35:45.018059779 +0000 UTC m=+662.536378697" lastFinishedPulling="2025-11-27 11:35:47.477949464 +0000 UTC m=+664.996268382" observedRunningTime="2025-11-27 11:35:48.065798475 +0000 UTC m=+665.584117393" watchObservedRunningTime="2025-11-27 11:35:48.068463738 +0000 UTC m=+665.586782646" Nov 27 11:35:48 crc kubenswrapper[4796]: I1127 11:35:48.089644 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-66p7v" podStartSLOduration=2.641582994 podStartE2EDuration="5.089622632s" podCreationTimestamp="2025-11-27 11:35:43 +0000 UTC" firstStartedPulling="2025-11-27 11:35:45.019196751 +0000 UTC m=+662.537515669" lastFinishedPulling="2025-11-27 11:35:47.467236389 +0000 UTC m=+664.985555307" observedRunningTime="2025-11-27 11:35:48.087173155 +0000 UTC m=+665.605492083" watchObservedRunningTime="2025-11-27 11:35:48.089622632 +0000 UTC m=+665.607941550" Nov 27 11:35:51 crc kubenswrapper[4796]: I1127 11:35:51.357571 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:51 crc kubenswrapper[4796]: I1127 11:35:51.357644 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:51 crc kubenswrapper[4796]: I1127 11:35:51.418375 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:51 crc kubenswrapper[4796]: I1127 11:35:51.559531 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:51 crc kubenswrapper[4796]: I1127 11:35:51.559566 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:51 crc kubenswrapper[4796]: I1127 11:35:51.608033 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:52 crc kubenswrapper[4796]: I1127 11:35:52.138448 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:35:52 crc kubenswrapper[4796]: I1127 11:35:52.141038 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:35:53 crc kubenswrapper[4796]: I1127 11:35:53.773543 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:53 crc kubenswrapper[4796]: I1127 11:35:53.774902 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:53 crc kubenswrapper[4796]: I1127 11:35:53.845072 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:54 crc kubenswrapper[4796]: I1127 11:35:54.003442 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:54 crc kubenswrapper[4796]: I1127 11:35:54.003491 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:54 crc kubenswrapper[4796]: I1127 11:35:54.062643 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:35:54 crc kubenswrapper[4796]: I1127 11:35:54.128076 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:35:54 crc kubenswrapper[4796]: I1127 11:35:54.130165 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:36:01 crc kubenswrapper[4796]: I1127 11:36:01.879762 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:36:01 crc kubenswrapper[4796]: I1127 11:36:01.880623 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:36:01 crc kubenswrapper[4796]: I1127 11:36:01.880717 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:36:01 crc kubenswrapper[4796]: I1127 11:36:01.881801 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39b36d5b8e5bafc99dea0a6943dc340119f3d41fbf1133a259918c397829bd29"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:36:01 crc kubenswrapper[4796]: I1127 11:36:01.882062 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://39b36d5b8e5bafc99dea0a6943dc340119f3d41fbf1133a259918c397829bd29" gracePeriod=600 Nov 27 11:36:02 crc kubenswrapper[4796]: I1127 11:36:02.135003 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="39b36d5b8e5bafc99dea0a6943dc340119f3d41fbf1133a259918c397829bd29" exitCode=0 Nov 27 11:36:02 crc kubenswrapper[4796]: I1127 11:36:02.135079 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"39b36d5b8e5bafc99dea0a6943dc340119f3d41fbf1133a259918c397829bd29"} Nov 27 11:36:02 crc kubenswrapper[4796]: I1127 11:36:02.135498 4796 scope.go:117] "RemoveContainer" containerID="2bc2f5728589faed16620248b20e0a440b5a67ed778b4753d5c8d2dca98cb558" Nov 27 11:36:03 crc kubenswrapper[4796]: I1127 11:36:03.147130 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"cfd84707341303f4e3c8588f23d09e0d87a32dbd7fe43f86a649e6799d5a18a9"} Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.392115 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sbxzr"] Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.393387 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.409566 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sbxzr"] Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499756 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-bound-sa-token\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499818 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499851 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l84m9\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-kube-api-access-l84m9\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499890 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fbc6bef9-fb41-4315-8e13-775b2f4ff435-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fbc6bef9-fb41-4315-8e13-775b2f4ff435-registry-certificates\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499937 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-registry-tls\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499959 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fbc6bef9-fb41-4315-8e13-775b2f4ff435-trusted-ca\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.499995 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fbc6bef9-fb41-4315-8e13-775b2f4ff435-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.540444 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601674 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-bound-sa-token\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601737 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l84m9\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-kube-api-access-l84m9\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601769 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fbc6bef9-fb41-4315-8e13-775b2f4ff435-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601793 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fbc6bef9-fb41-4315-8e13-775b2f4ff435-registry-certificates\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601823 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-registry-tls\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601844 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fbc6bef9-fb41-4315-8e13-775b2f4ff435-trusted-ca\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.601879 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fbc6bef9-fb41-4315-8e13-775b2f4ff435-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.603785 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fbc6bef9-fb41-4315-8e13-775b2f4ff435-trusted-ca\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.605100 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fbc6bef9-fb41-4315-8e13-775b2f4ff435-registry-certificates\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.605442 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fbc6bef9-fb41-4315-8e13-775b2f4ff435-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.607957 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fbc6bef9-fb41-4315-8e13-775b2f4ff435-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.608772 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-registry-tls\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.617027 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-bound-sa-token\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.620354 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l84m9\" (UniqueName: \"kubernetes.io/projected/fbc6bef9-fb41-4315-8e13-775b2f4ff435-kube-api-access-l84m9\") pod \"image-registry-66df7c8f76-sbxzr\" (UID: \"fbc6bef9-fb41-4315-8e13-775b2f4ff435\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.709138 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:08 crc kubenswrapper[4796]: I1127 11:37:08.894328 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sbxzr"] Nov 27 11:37:09 crc kubenswrapper[4796]: I1127 11:37:09.617905 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" event={"ID":"fbc6bef9-fb41-4315-8e13-775b2f4ff435","Type":"ContainerStarted","Data":"cf676282d6a0a1576aea16ee1aaa79acf65c9371ef31d1fa7ac6862507984df9"} Nov 27 11:37:09 crc kubenswrapper[4796]: I1127 11:37:09.618143 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" event={"ID":"fbc6bef9-fb41-4315-8e13-775b2f4ff435","Type":"ContainerStarted","Data":"1a432e7762dea434608f8a5e2278e4c60fbd17156e1acb1d40354c5c604c0f53"} Nov 27 11:37:09 crc kubenswrapper[4796]: I1127 11:37:09.618806 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:09 crc kubenswrapper[4796]: I1127 11:37:09.635411 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" podStartSLOduration=1.635393955 podStartE2EDuration="1.635393955s" podCreationTimestamp="2025-11-27 11:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:37:09.632903086 +0000 UTC m=+747.151222014" watchObservedRunningTime="2025-11-27 11:37:09.635393955 +0000 UTC m=+747.153712873" Nov 27 11:37:28 crc kubenswrapper[4796]: I1127 11:37:28.721612 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-sbxzr" Nov 27 11:37:28 crc kubenswrapper[4796]: I1127 11:37:28.781679 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5brmj"] Nov 27 11:37:40 crc kubenswrapper[4796]: I1127 11:37:40.861397 4796 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 11:37:53 crc kubenswrapper[4796]: I1127 11:37:53.832026 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" podUID="5e1b281d-0c21-49d8-89fd-ec2534405cdf" containerName="registry" containerID="cri-o://158891edaf4dfc616e75d3cd2bd662a321da935c0880abcc34f57648af671a4a" gracePeriod=30 Nov 27 11:37:53 crc kubenswrapper[4796]: I1127 11:37:53.981377 4796 generic.go:334] "Generic (PLEG): container finished" podID="5e1b281d-0c21-49d8-89fd-ec2534405cdf" containerID="158891edaf4dfc616e75d3cd2bd662a321da935c0880abcc34f57648af671a4a" exitCode=0 Nov 27 11:37:53 crc kubenswrapper[4796]: I1127 11:37:53.981445 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" event={"ID":"5e1b281d-0c21-49d8-89fd-ec2534405cdf","Type":"ContainerDied","Data":"158891edaf4dfc616e75d3cd2bd662a321da935c0880abcc34f57648af671a4a"} Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.232123 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.314187 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-trusted-ca\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.315072 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.315211 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.315509 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzjww\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-kube-api-access-xzjww\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.315533 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-tls\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.315549 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e1b281d-0c21-49d8-89fd-ec2534405cdf-installation-pull-secrets\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.315567 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-bound-sa-token\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.316196 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-certificates\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.316221 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e1b281d-0c21-49d8-89fd-ec2534405cdf-ca-trust-extracted\") pod \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\" (UID: \"5e1b281d-0c21-49d8-89fd-ec2534405cdf\") " Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.316465 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.317540 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.320951 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1b281d-0c21-49d8-89fd-ec2534405cdf-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.321150 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.324149 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-kube-api-access-xzjww" (OuterVolumeSpecName: "kube-api-access-xzjww") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "kube-api-access-xzjww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.325249 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.325560 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.342790 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1b281d-0c21-49d8-89fd-ec2534405cdf-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5e1b281d-0c21-49d8-89fd-ec2534405cdf" (UID: "5e1b281d-0c21-49d8-89fd-ec2534405cdf"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.417673 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzjww\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-kube-api-access-xzjww\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.417723 4796 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.417738 4796 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e1b281d-0c21-49d8-89fd-ec2534405cdf-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.417748 4796 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e1b281d-0c21-49d8-89fd-ec2534405cdf-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.417757 4796 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e1b281d-0c21-49d8-89fd-ec2534405cdf-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.417765 4796 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e1b281d-0c21-49d8-89fd-ec2534405cdf-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.993110 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" event={"ID":"5e1b281d-0c21-49d8-89fd-ec2534405cdf","Type":"ContainerDied","Data":"1f774e9f4c37dc25be8d05d91083f70b3e3d8d70d46ac376d7b5aaa666aedba1"} Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.993205 4796 scope.go:117] "RemoveContainer" containerID="158891edaf4dfc616e75d3cd2bd662a321da935c0880abcc34f57648af671a4a" Nov 27 11:37:54 crc kubenswrapper[4796]: I1127 11:37:54.993224 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5brmj" Nov 27 11:37:55 crc kubenswrapper[4796]: I1127 11:37:55.040231 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5brmj"] Nov 27 11:37:55 crc kubenswrapper[4796]: I1127 11:37:55.055172 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5brmj"] Nov 27 11:37:55 crc kubenswrapper[4796]: I1127 11:37:55.579804 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e1b281d-0c21-49d8-89fd-ec2534405cdf" path="/var/lib/kubelet/pods/5e1b281d-0c21-49d8-89fd-ec2534405cdf/volumes" Nov 27 11:38:31 crc kubenswrapper[4796]: I1127 11:38:31.879439 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:38:31 crc kubenswrapper[4796]: I1127 11:38:31.880021 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:39:01 crc kubenswrapper[4796]: I1127 11:39:01.879626 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:39:01 crc kubenswrapper[4796]: I1127 11:39:01.880312 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:39:31 crc kubenswrapper[4796]: I1127 11:39:31.879923 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:39:31 crc kubenswrapper[4796]: I1127 11:39:31.880595 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:39:31 crc kubenswrapper[4796]: I1127 11:39:31.880658 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:39:31 crc kubenswrapper[4796]: I1127 11:39:31.881485 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cfd84707341303f4e3c8588f23d09e0d87a32dbd7fe43f86a649e6799d5a18a9"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:39:31 crc kubenswrapper[4796]: I1127 11:39:31.881571 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://cfd84707341303f4e3c8588f23d09e0d87a32dbd7fe43f86a649e6799d5a18a9" gracePeriod=600 Nov 27 11:39:32 crc kubenswrapper[4796]: I1127 11:39:32.711656 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="cfd84707341303f4e3c8588f23d09e0d87a32dbd7fe43f86a649e6799d5a18a9" exitCode=0 Nov 27 11:39:32 crc kubenswrapper[4796]: I1127 11:39:32.711710 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"cfd84707341303f4e3c8588f23d09e0d87a32dbd7fe43f86a649e6799d5a18a9"} Nov 27 11:39:32 crc kubenswrapper[4796]: I1127 11:39:32.711917 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"b2bf926c9e1c204e1f7eb84bb17f79b027089b97e9942f906c34dda1a62f3cde"} Nov 27 11:39:32 crc kubenswrapper[4796]: I1127 11:39:32.711943 4796 scope.go:117] "RemoveContainer" containerID="39b36d5b8e5bafc99dea0a6943dc340119f3d41fbf1133a259918c397829bd29" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.950118 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nx48n"] Nov 27 11:41:08 crc kubenswrapper[4796]: E1127 11:41:08.951083 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1b281d-0c21-49d8-89fd-ec2534405cdf" containerName="registry" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.951104 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1b281d-0c21-49d8-89fd-ec2534405cdf" containerName="registry" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.951240 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e1b281d-0c21-49d8-89fd-ec2534405cdf" containerName="registry" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.951724 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nx48n" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.954929 4796 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-nhn9t" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.955355 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.955628 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.955822 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-scllt"] Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.956787 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.963540 4796 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-lzkmk" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.963936 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-scllt"] Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.968406 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nx48n"] Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.975490 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qrd89"] Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.976351 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.980015 4796 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gkhtz" Nov 27 11:41:08 crc kubenswrapper[4796]: I1127 11:41:08.993497 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qrd89"] Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.055276 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpffl\" (UniqueName: \"kubernetes.io/projected/bfed3074-7dee-4f77-900a-8546651ce2e6-kube-api-access-kpffl\") pod \"cert-manager-webhook-5655c58dd6-qrd89\" (UID: \"bfed3074-7dee-4f77-900a-8546651ce2e6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.055333 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cck8\" (UniqueName: \"kubernetes.io/projected/2a72ae09-3cbb-4ab3-91b3-0aab83b57b74-kube-api-access-6cck8\") pod \"cert-manager-cainjector-7f985d654d-scllt\" (UID: \"2a72ae09-3cbb-4ab3-91b3-0aab83b57b74\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.055394 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c5j4\" (UniqueName: \"kubernetes.io/projected/22e38aa1-4d12-4b1f-bfef-12d303bc12ca-kube-api-access-5c5j4\") pod \"cert-manager-5b446d88c5-nx48n\" (UID: \"22e38aa1-4d12-4b1f-bfef-12d303bc12ca\") " pod="cert-manager/cert-manager-5b446d88c5-nx48n" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.156125 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c5j4\" (UniqueName: \"kubernetes.io/projected/22e38aa1-4d12-4b1f-bfef-12d303bc12ca-kube-api-access-5c5j4\") pod \"cert-manager-5b446d88c5-nx48n\" (UID: \"22e38aa1-4d12-4b1f-bfef-12d303bc12ca\") " pod="cert-manager/cert-manager-5b446d88c5-nx48n" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.156200 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpffl\" (UniqueName: \"kubernetes.io/projected/bfed3074-7dee-4f77-900a-8546651ce2e6-kube-api-access-kpffl\") pod \"cert-manager-webhook-5655c58dd6-qrd89\" (UID: \"bfed3074-7dee-4f77-900a-8546651ce2e6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.156235 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cck8\" (UniqueName: \"kubernetes.io/projected/2a72ae09-3cbb-4ab3-91b3-0aab83b57b74-kube-api-access-6cck8\") pod \"cert-manager-cainjector-7f985d654d-scllt\" (UID: \"2a72ae09-3cbb-4ab3-91b3-0aab83b57b74\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.175525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpffl\" (UniqueName: \"kubernetes.io/projected/bfed3074-7dee-4f77-900a-8546651ce2e6-kube-api-access-kpffl\") pod \"cert-manager-webhook-5655c58dd6-qrd89\" (UID: \"bfed3074-7dee-4f77-900a-8546651ce2e6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.175603 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cck8\" (UniqueName: \"kubernetes.io/projected/2a72ae09-3cbb-4ab3-91b3-0aab83b57b74-kube-api-access-6cck8\") pod \"cert-manager-cainjector-7f985d654d-scllt\" (UID: \"2a72ae09-3cbb-4ab3-91b3-0aab83b57b74\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.177096 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c5j4\" (UniqueName: \"kubernetes.io/projected/22e38aa1-4d12-4b1f-bfef-12d303bc12ca-kube-api-access-5c5j4\") pod \"cert-manager-5b446d88c5-nx48n\" (UID: \"22e38aa1-4d12-4b1f-bfef-12d303bc12ca\") " pod="cert-manager/cert-manager-5b446d88c5-nx48n" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.278063 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nx48n" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.289855 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.295939 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.514233 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nx48n"] Nov 27 11:41:09 crc kubenswrapper[4796]: W1127 11:41:09.520448 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22e38aa1_4d12_4b1f_bfef_12d303bc12ca.slice/crio-a3fdb747ad2aa99b84fbeb98cd31671ace6886e39e498525f934af817316c3a1 WatchSource:0}: Error finding container a3fdb747ad2aa99b84fbeb98cd31671ace6886e39e498525f934af817316c3a1: Status 404 returned error can't find the container with id a3fdb747ad2aa99b84fbeb98cd31671ace6886e39e498525f934af817316c3a1 Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.524014 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.778164 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-scllt"] Nov 27 11:41:09 crc kubenswrapper[4796]: W1127 11:41:09.781711 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a72ae09_3cbb_4ab3_91b3_0aab83b57b74.slice/crio-c9fe6645fa82fe25d61cc834cce809c2195373bab1deeffc84248b737b9dbce2 WatchSource:0}: Error finding container c9fe6645fa82fe25d61cc834cce809c2195373bab1deeffc84248b737b9dbce2: Status 404 returned error can't find the container with id c9fe6645fa82fe25d61cc834cce809c2195373bab1deeffc84248b737b9dbce2 Nov 27 11:41:09 crc kubenswrapper[4796]: I1127 11:41:09.800823 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qrd89"] Nov 27 11:41:09 crc kubenswrapper[4796]: W1127 11:41:09.805684 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfed3074_7dee_4f77_900a_8546651ce2e6.slice/crio-d99b9d5e23ab67a873da487c1fe948b89703656c7e3974a498366f72e44b03e8 WatchSource:0}: Error finding container d99b9d5e23ab67a873da487c1fe948b89703656c7e3974a498366f72e44b03e8: Status 404 returned error can't find the container with id d99b9d5e23ab67a873da487c1fe948b89703656c7e3974a498366f72e44b03e8 Nov 27 11:41:10 crc kubenswrapper[4796]: I1127 11:41:10.354696 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nx48n" event={"ID":"22e38aa1-4d12-4b1f-bfef-12d303bc12ca","Type":"ContainerStarted","Data":"a3fdb747ad2aa99b84fbeb98cd31671ace6886e39e498525f934af817316c3a1"} Nov 27 11:41:10 crc kubenswrapper[4796]: I1127 11:41:10.356029 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" event={"ID":"bfed3074-7dee-4f77-900a-8546651ce2e6","Type":"ContainerStarted","Data":"d99b9d5e23ab67a873da487c1fe948b89703656c7e3974a498366f72e44b03e8"} Nov 27 11:41:10 crc kubenswrapper[4796]: I1127 11:41:10.357114 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" event={"ID":"2a72ae09-3cbb-4ab3-91b3-0aab83b57b74","Type":"ContainerStarted","Data":"c9fe6645fa82fe25d61cc834cce809c2195373bab1deeffc84248b737b9dbce2"} Nov 27 11:41:12 crc kubenswrapper[4796]: I1127 11:41:12.371525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nx48n" event={"ID":"22e38aa1-4d12-4b1f-bfef-12d303bc12ca","Type":"ContainerStarted","Data":"2c869711db83c6d5e402f4cc6d956ceb91e3477d0b1f48f9f9f17eada92dac5b"} Nov 27 11:41:12 crc kubenswrapper[4796]: I1127 11:41:12.388081 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nx48n" podStartSLOduration=2.276051251 podStartE2EDuration="4.388060015s" podCreationTimestamp="2025-11-27 11:41:08 +0000 UTC" firstStartedPulling="2025-11-27 11:41:09.523801913 +0000 UTC m=+987.042120831" lastFinishedPulling="2025-11-27 11:41:11.635810657 +0000 UTC m=+989.154129595" observedRunningTime="2025-11-27 11:41:12.386209655 +0000 UTC m=+989.904528583" watchObservedRunningTime="2025-11-27 11:41:12.388060015 +0000 UTC m=+989.906378943" Nov 27 11:41:13 crc kubenswrapper[4796]: I1127 11:41:13.377980 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" event={"ID":"bfed3074-7dee-4f77-900a-8546651ce2e6","Type":"ContainerStarted","Data":"09cdd09a107598638b2b3bfccaaabc78b998f112c6d9c8f9a88b9fe6bab93649"} Nov 27 11:41:13 crc kubenswrapper[4796]: I1127 11:41:13.378084 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:13 crc kubenswrapper[4796]: I1127 11:41:13.379769 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" event={"ID":"2a72ae09-3cbb-4ab3-91b3-0aab83b57b74","Type":"ContainerStarted","Data":"f38ff6ba4ec214e2a8dfad89687eeb78c26a024151d4bf2f4e95ca63077088c7"} Nov 27 11:41:13 crc kubenswrapper[4796]: I1127 11:41:13.399771 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" podStartSLOduration=2.423135717 podStartE2EDuration="5.399753394s" podCreationTimestamp="2025-11-27 11:41:08 +0000 UTC" firstStartedPulling="2025-11-27 11:41:09.808810314 +0000 UTC m=+987.327129232" lastFinishedPulling="2025-11-27 11:41:12.785427991 +0000 UTC m=+990.303746909" observedRunningTime="2025-11-27 11:41:13.395330334 +0000 UTC m=+990.913649282" watchObservedRunningTime="2025-11-27 11:41:13.399753394 +0000 UTC m=+990.918072322" Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.300785 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-qrd89" Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.318612 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-scllt" podStartSLOduration=8.248718797 podStartE2EDuration="11.318583747s" podCreationTimestamp="2025-11-27 11:41:08 +0000 UTC" firstStartedPulling="2025-11-27 11:41:09.784492988 +0000 UTC m=+987.302811896" lastFinishedPulling="2025-11-27 11:41:12.854357938 +0000 UTC m=+990.372676846" observedRunningTime="2025-11-27 11:41:13.412688309 +0000 UTC m=+990.931007247" watchObservedRunningTime="2025-11-27 11:41:19.318583747 +0000 UTC m=+996.836902665" Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.410722 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6vnft"] Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.411479 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-controller" containerID="cri-o://d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.411867 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="nbdb" containerID="cri-o://d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.412018 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-acl-logging" containerID="cri-o://f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.412171 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="sbdb" containerID="cri-o://84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.412209 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-node" containerID="cri-o://1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.412357 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="northd" containerID="cri-o://92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.412670 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" gracePeriod=30 Nov 27 11:41:19 crc kubenswrapper[4796]: I1127 11:41:19.444500 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" containerID="cri-o://84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" gracePeriod=30 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.162394 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/3.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.165904 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovn-acl-logging/0.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.166496 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovn-controller/0.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.166948 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.205764 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-script-lib\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206075 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-netd\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206209 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206420 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovn-node-metrics-cert\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206602 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-ovn-kubernetes\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206784 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-slash\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206440 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206695 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.206922 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-slash" (OuterVolumeSpecName: "host-slash") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207049 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-env-overrides\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207169 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-openvswitch\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207195 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-systemd\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207216 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-node-log\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207242 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-bin\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207249 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207317 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-var-lib-openvswitch\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207347 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207374 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-node-log" (OuterVolumeSpecName: "node-log") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207409 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-etc-openvswitch\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207422 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207447 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207469 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-systemd-units\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207516 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-netns\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207517 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207566 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-config\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207616 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207573 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207678 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-kubelet\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207693 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207726 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgvrx\" (UniqueName: \"kubernetes.io/projected/9cef17d7-9643-44ec-ae40-cc46b875ed42-kube-api-access-pgvrx\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207770 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-log-socket\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207737 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.207819 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-ovn\") pod \"9cef17d7-9643-44ec-ae40-cc46b875ed42\" (UID: \"9cef17d7-9643-44ec-ae40-cc46b875ed42\") " Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208367 4796 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208404 4796 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-node-log\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208429 4796 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208456 4796 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208461 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208480 4796 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208506 4796 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208521 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-log-socket" (OuterVolumeSpecName: "log-socket") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208530 4796 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208580 4796 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208607 4796 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208629 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208649 4796 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208666 4796 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208684 4796 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-host-slash\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.208579 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.210153 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.213151 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cef17d7-9643-44ec-ae40-cc46b875ed42-kube-api-access-pgvrx" (OuterVolumeSpecName: "kube-api-access-pgvrx") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "kube-api-access-pgvrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.213663 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.234340 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9cef17d7-9643-44ec-ae40-cc46b875ed42" (UID: "9cef17d7-9643-44ec-ae40-cc46b875ed42"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.243800 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9xxqp"] Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245577 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-node" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245621 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-node" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245643 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="nbdb" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245655 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="nbdb" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245669 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245682 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245696 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245709 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245723 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245735 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245752 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="sbdb" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245765 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="sbdb" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245783 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245797 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245811 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-acl-logging" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.245933 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-acl-logging" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.245988 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.246007 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.246031 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="northd" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247067 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="northd" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.247104 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kubecfg-setup" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247121 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kubecfg-setup" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.247143 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247165 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247444 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="nbdb" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247476 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247499 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247520 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovn-acl-logging" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247545 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247563 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247583 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="sbdb" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247603 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247619 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-node" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247643 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.247665 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="northd" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.248032 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.248057 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.248294 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerName="ovnkube-controller" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.251918 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309714 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-cni-bin\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309772 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovn-node-metrics-cert\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309801 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-slash\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-var-lib-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309849 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv927\" (UniqueName: \"kubernetes.io/projected/113a1a41-30b9-4fa4-924b-0ce363bdb289-kube-api-access-sv927\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309872 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-run-ovn-kubernetes\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309895 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-systemd\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309921 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-etc-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309948 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309968 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-systemd-units\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.309984 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310006 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-env-overrides\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310026 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-kubelet\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310047 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovnkube-script-lib\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310066 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-ovn\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310096 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-run-netns\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310136 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-cni-netd\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310153 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-log-socket\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310174 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovnkube-config\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-node-log\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310241 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310255 4796 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310290 4796 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310303 4796 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9cef17d7-9643-44ec-ae40-cc46b875ed42-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310315 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgvrx\" (UniqueName: \"kubernetes.io/projected/9cef17d7-9643-44ec-ae40-cc46b875ed42-kube-api-access-pgvrx\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310325 4796 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-log-socket\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.310335 4796 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9cef17d7-9643-44ec-ae40-cc46b875ed42-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411457 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv927\" (UniqueName: \"kubernetes.io/projected/113a1a41-30b9-4fa4-924b-0ce363bdb289-kube-api-access-sv927\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411526 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-run-ovn-kubernetes\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411567 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-systemd\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411590 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-run-ovn-kubernetes\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411606 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-etc-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411658 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-etc-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411655 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411714 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-systemd-units\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411684 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411776 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-env-overrides\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411783 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-systemd-units\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411737 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-systemd\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411807 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-kubelet\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411838 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-kubelet\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovnkube-script-lib\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.411864 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412013 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-ovn\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412128 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-run-netns\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412137 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-run-ovn\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412187 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-run-netns\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412222 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-log-socket\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412243 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-log-socket\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412255 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-cni-netd\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412308 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-cni-netd\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovnkube-config\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412373 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-node-log\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412414 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-cni-bin\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412467 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovn-node-metrics-cert\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412500 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-slash\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412539 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-var-lib-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412629 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-var-lib-openvswitch\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412687 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-node-log\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412748 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-cni-bin\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412962 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovnkube-script-lib\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.412996 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-env-overrides\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.413113 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/113a1a41-30b9-4fa4-924b-0ce363bdb289-host-slash\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.413131 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovnkube-config\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.418414 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/113a1a41-30b9-4fa4-924b-0ce363bdb289-ovn-node-metrics-cert\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.437363 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv927\" (UniqueName: \"kubernetes.io/projected/113a1a41-30b9-4fa4-924b-0ce363bdb289-kube-api-access-sv927\") pod \"ovnkube-node-9xxqp\" (UID: \"113a1a41-30b9-4fa4-924b-0ce363bdb289\") " pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.442676 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/2.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.444085 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/1.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.444178 4796 generic.go:334] "Generic (PLEG): container finished" podID="1f8355a9-d2c1-4c54-be9d-68ef66397560" containerID="12e4f74214e01b5188063377438e0ba34315407131a3a43c591387bc0018e91b" exitCode=2 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.444373 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerDied","Data":"12e4f74214e01b5188063377438e0ba34315407131a3a43c591387bc0018e91b"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.444455 4796 scope.go:117] "RemoveContainer" containerID="6c486fe7646d0aa2b3c40a37709989d6bba63597245b72fe4f617a4ced2cf70b" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.444957 4796 scope.go:117] "RemoveContainer" containerID="12e4f74214e01b5188063377438e0ba34315407131a3a43c591387bc0018e91b" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.447638 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovnkube-controller/3.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.453701 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovn-acl-logging/0.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.454647 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6vnft_9cef17d7-9643-44ec-ae40-cc46b875ed42/ovn-controller/0.log" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455099 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" exitCode=0 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455138 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" exitCode=0 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455154 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" exitCode=0 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455170 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" exitCode=0 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455189 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" exitCode=0 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455203 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" exitCode=0 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455216 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" exitCode=143 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455236 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cef17d7-9643-44ec-ae40-cc46b875ed42" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" exitCode=143 Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455237 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455218 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455320 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455346 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455371 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455410 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455441 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455458 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455470 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455481 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455492 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455504 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455515 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455526 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455537 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455547 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455562 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455578 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455591 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455602 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455613 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455623 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455634 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455645 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455655 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455667 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455678 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455693 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455708 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455720 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455732 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455743 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455753 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455764 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455774 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455801 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455812 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455824 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455843 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6vnft" event={"ID":"9cef17d7-9643-44ec-ae40-cc46b875ed42","Type":"ContainerDied","Data":"ef1f4ab0b94528f62ddeb26f7105cb62b5c07150876b5e695500d256cd136371"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455863 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455879 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455893 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455907 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455918 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455928 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455939 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455950 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455961 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.455972 4796 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.489067 4796 scope.go:117] "RemoveContainer" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.510065 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6vnft"] Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.513542 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6vnft"] Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.526567 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.571477 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.597449 4796 scope.go:117] "RemoveContainer" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" Nov 27 11:41:20 crc kubenswrapper[4796]: W1127 11:41:20.612235 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod113a1a41_30b9_4fa4_924b_0ce363bdb289.slice/crio-fd458071bf4ce02ff41093854e7472efe0e88c46eea068079747c3400eb74d4d WatchSource:0}: Error finding container fd458071bf4ce02ff41093854e7472efe0e88c46eea068079747c3400eb74d4d: Status 404 returned error can't find the container with id fd458071bf4ce02ff41093854e7472efe0e88c46eea068079747c3400eb74d4d Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.684708 4796 scope.go:117] "RemoveContainer" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.702754 4796 scope.go:117] "RemoveContainer" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.738895 4796 scope.go:117] "RemoveContainer" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.759763 4796 scope.go:117] "RemoveContainer" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.776327 4796 scope.go:117] "RemoveContainer" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.862890 4796 scope.go:117] "RemoveContainer" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.884656 4796 scope.go:117] "RemoveContainer" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.913547 4796 scope.go:117] "RemoveContainer" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.914093 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": container with ID starting with 84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e not found: ID does not exist" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.914141 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} err="failed to get container status \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": rpc error: code = NotFound desc = could not find container \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": container with ID starting with 84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.914177 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.918301 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": container with ID starting with fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4 not found: ID does not exist" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.918347 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} err="failed to get container status \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": rpc error: code = NotFound desc = could not find container \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": container with ID starting with fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.918374 4796 scope.go:117] "RemoveContainer" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.918781 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": container with ID starting with 84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79 not found: ID does not exist" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.918810 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} err="failed to get container status \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": rpc error: code = NotFound desc = could not find container \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": container with ID starting with 84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.918828 4796 scope.go:117] "RemoveContainer" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.919306 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": container with ID starting with d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2 not found: ID does not exist" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.919337 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} err="failed to get container status \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": rpc error: code = NotFound desc = could not find container \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": container with ID starting with d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.919363 4796 scope.go:117] "RemoveContainer" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.919785 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": container with ID starting with 92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1 not found: ID does not exist" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.919808 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} err="failed to get container status \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": rpc error: code = NotFound desc = could not find container \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": container with ID starting with 92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.919824 4796 scope.go:117] "RemoveContainer" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.920093 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": container with ID starting with 935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38 not found: ID does not exist" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.920121 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} err="failed to get container status \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": rpc error: code = NotFound desc = could not find container \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": container with ID starting with 935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.920138 4796 scope.go:117] "RemoveContainer" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.920603 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": container with ID starting with 1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680 not found: ID does not exist" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.920639 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} err="failed to get container status \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": rpc error: code = NotFound desc = could not find container \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": container with ID starting with 1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.920659 4796 scope.go:117] "RemoveContainer" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.920884 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": container with ID starting with f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903 not found: ID does not exist" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.920913 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} err="failed to get container status \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": rpc error: code = NotFound desc = could not find container \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": container with ID starting with f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.920942 4796 scope.go:117] "RemoveContainer" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.921150 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": container with ID starting with d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944 not found: ID does not exist" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921176 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} err="failed to get container status \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": rpc error: code = NotFound desc = could not find container \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": container with ID starting with d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921192 4796 scope.go:117] "RemoveContainer" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" Nov 27 11:41:20 crc kubenswrapper[4796]: E1127 11:41:20.921489 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": container with ID starting with daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74 not found: ID does not exist" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921519 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} err="failed to get container status \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": rpc error: code = NotFound desc = could not find container \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": container with ID starting with daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921535 4796 scope.go:117] "RemoveContainer" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921760 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} err="failed to get container status \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": rpc error: code = NotFound desc = could not find container \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": container with ID starting with 84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921783 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.921990 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} err="failed to get container status \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": rpc error: code = NotFound desc = could not find container \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": container with ID starting with fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922014 4796 scope.go:117] "RemoveContainer" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922200 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} err="failed to get container status \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": rpc error: code = NotFound desc = could not find container \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": container with ID starting with 84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922221 4796 scope.go:117] "RemoveContainer" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922415 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} err="failed to get container status \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": rpc error: code = NotFound desc = could not find container \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": container with ID starting with d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922439 4796 scope.go:117] "RemoveContainer" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922646 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} err="failed to get container status \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": rpc error: code = NotFound desc = could not find container \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": container with ID starting with 92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922667 4796 scope.go:117] "RemoveContainer" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922859 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} err="failed to get container status \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": rpc error: code = NotFound desc = could not find container \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": container with ID starting with 935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.922881 4796 scope.go:117] "RemoveContainer" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923066 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} err="failed to get container status \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": rpc error: code = NotFound desc = could not find container \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": container with ID starting with 1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923089 4796 scope.go:117] "RemoveContainer" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923311 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} err="failed to get container status \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": rpc error: code = NotFound desc = could not find container \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": container with ID starting with f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923334 4796 scope.go:117] "RemoveContainer" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923540 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} err="failed to get container status \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": rpc error: code = NotFound desc = could not find container \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": container with ID starting with d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923563 4796 scope.go:117] "RemoveContainer" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923747 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} err="failed to get container status \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": rpc error: code = NotFound desc = could not find container \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": container with ID starting with daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923769 4796 scope.go:117] "RemoveContainer" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923962 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} err="failed to get container status \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": rpc error: code = NotFound desc = could not find container \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": container with ID starting with 84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.923983 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924155 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} err="failed to get container status \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": rpc error: code = NotFound desc = could not find container \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": container with ID starting with fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924177 4796 scope.go:117] "RemoveContainer" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924432 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} err="failed to get container status \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": rpc error: code = NotFound desc = could not find container \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": container with ID starting with 84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924454 4796 scope.go:117] "RemoveContainer" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924654 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} err="failed to get container status \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": rpc error: code = NotFound desc = could not find container \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": container with ID starting with d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924675 4796 scope.go:117] "RemoveContainer" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924857 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} err="failed to get container status \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": rpc error: code = NotFound desc = could not find container \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": container with ID starting with 92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.924878 4796 scope.go:117] "RemoveContainer" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925079 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} err="failed to get container status \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": rpc error: code = NotFound desc = could not find container \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": container with ID starting with 935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925099 4796 scope.go:117] "RemoveContainer" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925317 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} err="failed to get container status \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": rpc error: code = NotFound desc = could not find container \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": container with ID starting with 1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925339 4796 scope.go:117] "RemoveContainer" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925567 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} err="failed to get container status \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": rpc error: code = NotFound desc = could not find container \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": container with ID starting with f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925588 4796 scope.go:117] "RemoveContainer" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925783 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} err="failed to get container status \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": rpc error: code = NotFound desc = could not find container \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": container with ID starting with d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.925804 4796 scope.go:117] "RemoveContainer" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926003 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} err="failed to get container status \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": rpc error: code = NotFound desc = could not find container \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": container with ID starting with daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926024 4796 scope.go:117] "RemoveContainer" containerID="84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926200 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e"} err="failed to get container status \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": rpc error: code = NotFound desc = could not find container \"84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e\": container with ID starting with 84271d6ade7a4b3e7a953c81cddb8ccb2389f96cee3f9a00f140c0febf59788e not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926221 4796 scope.go:117] "RemoveContainer" containerID="fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926455 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4"} err="failed to get container status \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": rpc error: code = NotFound desc = could not find container \"fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4\": container with ID starting with fec98994af11626c6553fe1d1c1f685e82b88ff5bb701ceb3bdc38dd82f1d6a4 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926479 4796 scope.go:117] "RemoveContainer" containerID="84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926713 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79"} err="failed to get container status \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": rpc error: code = NotFound desc = could not find container \"84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79\": container with ID starting with 84357e0e1144352a6012a74f3c6db411a57c4efd14cfd8f07a81a1ce8711fe79 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926738 4796 scope.go:117] "RemoveContainer" containerID="d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.926982 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2"} err="failed to get container status \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": rpc error: code = NotFound desc = could not find container \"d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2\": container with ID starting with d9d1822a3e84b3c47ede305c8e4524aa100e226fc11a54a4d75e56cd5760d0b2 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927003 4796 scope.go:117] "RemoveContainer" containerID="92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927243 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1"} err="failed to get container status \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": rpc error: code = NotFound desc = could not find container \"92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1\": container with ID starting with 92f372c19d14df653981ddc3c68498b806bffb554931e1ea26b764f265d9ccb1 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927284 4796 scope.go:117] "RemoveContainer" containerID="935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927548 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38"} err="failed to get container status \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": rpc error: code = NotFound desc = could not find container \"935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38\": container with ID starting with 935525c6ecd486a5a895c35c0aa35fc436280c332749d0da550a7f2e3b5ecc38 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927570 4796 scope.go:117] "RemoveContainer" containerID="1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927930 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680"} err="failed to get container status \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": rpc error: code = NotFound desc = could not find container \"1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680\": container with ID starting with 1cce80395910950fd27e7e5eb00315f90c3dc345724ac3cd7db8de38a0778680 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.927956 4796 scope.go:117] "RemoveContainer" containerID="f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.928211 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903"} err="failed to get container status \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": rpc error: code = NotFound desc = could not find container \"f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903\": container with ID starting with f8fe1c4fca5042605525c98268e5bafa79836682cecdd8e159544c40ba31e903 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.928234 4796 scope.go:117] "RemoveContainer" containerID="d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.928634 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944"} err="failed to get container status \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": rpc error: code = NotFound desc = could not find container \"d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944\": container with ID starting with d4d65f0de632b9a492013a0ec1426951a55f8c4f48a9d55dea2dac98faaa9944 not found: ID does not exist" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.928659 4796 scope.go:117] "RemoveContainer" containerID="daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74" Nov 27 11:41:20 crc kubenswrapper[4796]: I1127 11:41:20.928936 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74"} err="failed to get container status \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": rpc error: code = NotFound desc = could not find container \"daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74\": container with ID starting with daf49a87a633f081844800e0654264869983c62c4f227a2cd56eb62f3e95cb74 not found: ID does not exist" Nov 27 11:41:21 crc kubenswrapper[4796]: I1127 11:41:21.465577 4796 generic.go:334] "Generic (PLEG): container finished" podID="113a1a41-30b9-4fa4-924b-0ce363bdb289" containerID="0046aac4a8470efc49cebaa7ffcf31efbf975e908dc72d9ceee4990098d6be02" exitCode=0 Nov 27 11:41:21 crc kubenswrapper[4796]: I1127 11:41:21.465654 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerDied","Data":"0046aac4a8470efc49cebaa7ffcf31efbf975e908dc72d9ceee4990098d6be02"} Nov 27 11:41:21 crc kubenswrapper[4796]: I1127 11:41:21.466050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"fd458071bf4ce02ff41093854e7472efe0e88c46eea068079747c3400eb74d4d"} Nov 27 11:41:21 crc kubenswrapper[4796]: I1127 11:41:21.469114 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8drhq_1f8355a9-d2c1-4c54-be9d-68ef66397560/kube-multus/2.log" Nov 27 11:41:21 crc kubenswrapper[4796]: I1127 11:41:21.469176 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8drhq" event={"ID":"1f8355a9-d2c1-4c54-be9d-68ef66397560","Type":"ContainerStarted","Data":"d945310fec954e400ff0affbe66f817c30609366ee2fb55618ea96597b690137"} Nov 27 11:41:21 crc kubenswrapper[4796]: I1127 11:41:21.582532 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cef17d7-9643-44ec-ae40-cc46b875ed42" path="/var/lib/kubelet/pods/9cef17d7-9643-44ec-ae40-cc46b875ed42/volumes" Nov 27 11:41:22 crc kubenswrapper[4796]: I1127 11:41:22.490818 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"1c48e02f45809779c804e118cfc7ff56e395abe1a93da4af1cc51ecfb661dc4c"} Nov 27 11:41:22 crc kubenswrapper[4796]: I1127 11:41:22.491181 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"5874e5d92f276baae876368227acff3799fd6ed57549448100f4747055673cb9"} Nov 27 11:41:22 crc kubenswrapper[4796]: I1127 11:41:22.491192 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"73463d64d1586cf733cdf3c3852c69ea7e09535d925ce8bcefa6a29d133d3777"} Nov 27 11:41:22 crc kubenswrapper[4796]: I1127 11:41:22.491202 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"af5aab39be556749e832a9c0f4f6dd6f6bddd8c8d5c760664e4cc661b840c9b4"} Nov 27 11:41:22 crc kubenswrapper[4796]: I1127 11:41:22.491210 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"f617eb6bcf34ff6497624cafc74727ce52c8f141df90b44ffecc75948a74cf2e"} Nov 27 11:41:23 crc kubenswrapper[4796]: I1127 11:41:23.506648 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"2b65bf6d38c77ef0489da36ae1723c9ae61831ffc7cc65807989b62beb28f225"} Nov 27 11:41:25 crc kubenswrapper[4796]: I1127 11:41:25.524909 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"f7970dc0e3fdc674d824d6776487e2f696c8630f01277cf50a5a3917cf3fce59"} Nov 27 11:41:27 crc kubenswrapper[4796]: I1127 11:41:27.549900 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" event={"ID":"113a1a41-30b9-4fa4-924b-0ce363bdb289","Type":"ContainerStarted","Data":"bb084bbfe9673555437baecf93ae47432d27940a5a214890dba3b9709edc9b1b"} Nov 27 11:41:27 crc kubenswrapper[4796]: I1127 11:41:27.550648 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:27 crc kubenswrapper[4796]: I1127 11:41:27.550669 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:27 crc kubenswrapper[4796]: I1127 11:41:27.578927 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:27 crc kubenswrapper[4796]: I1127 11:41:27.588952 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" podStartSLOduration=7.588907589 podStartE2EDuration="7.588907589s" podCreationTimestamp="2025-11-27 11:41:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:41:27.581031573 +0000 UTC m=+1005.099350481" watchObservedRunningTime="2025-11-27 11:41:27.588907589 +0000 UTC m=+1005.107226517" Nov 27 11:41:28 crc kubenswrapper[4796]: I1127 11:41:28.559596 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:28 crc kubenswrapper[4796]: I1127 11:41:28.624140 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.751124 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-75kwt"] Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.755235 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.769993 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-75kwt"] Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.892327 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqbq5\" (UniqueName: \"kubernetes.io/projected/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-kube-api-access-gqbq5\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.892385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-utilities\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.892422 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-catalog-content\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.994078 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-catalog-content\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.994189 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqbq5\" (UniqueName: \"kubernetes.io/projected/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-kube-api-access-gqbq5\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.994223 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-utilities\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.994731 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-utilities\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:34 crc kubenswrapper[4796]: I1127 11:41:34.995044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-catalog-content\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:35 crc kubenswrapper[4796]: I1127 11:41:35.017489 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqbq5\" (UniqueName: \"kubernetes.io/projected/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-kube-api-access-gqbq5\") pod \"community-operators-75kwt\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:35 crc kubenswrapper[4796]: I1127 11:41:35.135728 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:35 crc kubenswrapper[4796]: I1127 11:41:35.391415 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-75kwt"] Nov 27 11:41:35 crc kubenswrapper[4796]: W1127 11:41:35.395739 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb6dd66e_7ccf_417e_9ae8_f837eee4cdce.slice/crio-7594686ea46f74a507e697443e0760396664c3e60960acb35216fed9de0e982d WatchSource:0}: Error finding container 7594686ea46f74a507e697443e0760396664c3e60960acb35216fed9de0e982d: Status 404 returned error can't find the container with id 7594686ea46f74a507e697443e0760396664c3e60960acb35216fed9de0e982d Nov 27 11:41:35 crc kubenswrapper[4796]: I1127 11:41:35.607462 4796 generic.go:334] "Generic (PLEG): container finished" podID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerID="33f0fb1cea4f41dbd8dff0f0fd8216566098853cc8343ea202a69c9cfe5cd765" exitCode=0 Nov 27 11:41:35 crc kubenswrapper[4796]: I1127 11:41:35.607523 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerDied","Data":"33f0fb1cea4f41dbd8dff0f0fd8216566098853cc8343ea202a69c9cfe5cd765"} Nov 27 11:41:35 crc kubenswrapper[4796]: I1127 11:41:35.607770 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerStarted","Data":"7594686ea46f74a507e697443e0760396664c3e60960acb35216fed9de0e982d"} Nov 27 11:41:36 crc kubenswrapper[4796]: I1127 11:41:36.618061 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerStarted","Data":"c6ca6694c403573d67ef123ce7f86f6aad2782355bc0d96e17a133a9dab7f010"} Nov 27 11:41:37 crc kubenswrapper[4796]: I1127 11:41:37.625380 4796 generic.go:334] "Generic (PLEG): container finished" podID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerID="c6ca6694c403573d67ef123ce7f86f6aad2782355bc0d96e17a133a9dab7f010" exitCode=0 Nov 27 11:41:37 crc kubenswrapper[4796]: I1127 11:41:37.625433 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerDied","Data":"c6ca6694c403573d67ef123ce7f86f6aad2782355bc0d96e17a133a9dab7f010"} Nov 27 11:41:38 crc kubenswrapper[4796]: I1127 11:41:38.638467 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerStarted","Data":"7e759e73b228715e0cb9bfe35cc7ff831cafb0dcc2f69a7b17b5d14ae199ee6e"} Nov 27 11:41:45 crc kubenswrapper[4796]: I1127 11:41:45.136829 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:45 crc kubenswrapper[4796]: I1127 11:41:45.137536 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:45 crc kubenswrapper[4796]: I1127 11:41:45.215485 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:45 crc kubenswrapper[4796]: I1127 11:41:45.237659 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-75kwt" podStartSLOduration=8.617543187 podStartE2EDuration="11.237639209s" podCreationTimestamp="2025-11-27 11:41:34 +0000 UTC" firstStartedPulling="2025-11-27 11:41:35.608681592 +0000 UTC m=+1013.127000520" lastFinishedPulling="2025-11-27 11:41:38.228777624 +0000 UTC m=+1015.747096542" observedRunningTime="2025-11-27 11:41:38.665411703 +0000 UTC m=+1016.183730621" watchObservedRunningTime="2025-11-27 11:41:45.237639209 +0000 UTC m=+1022.755958137" Nov 27 11:41:45 crc kubenswrapper[4796]: I1127 11:41:45.733080 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:45 crc kubenswrapper[4796]: I1127 11:41:45.778909 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-75kwt"] Nov 27 11:41:47 crc kubenswrapper[4796]: I1127 11:41:47.693118 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-75kwt" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="registry-server" containerID="cri-o://7e759e73b228715e0cb9bfe35cc7ff831cafb0dcc2f69a7b17b5d14ae199ee6e" gracePeriod=2 Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.714202 4796 generic.go:334] "Generic (PLEG): container finished" podID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerID="7e759e73b228715e0cb9bfe35cc7ff831cafb0dcc2f69a7b17b5d14ae199ee6e" exitCode=0 Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.714348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerDied","Data":"7e759e73b228715e0cb9bfe35cc7ff831cafb0dcc2f69a7b17b5d14ae199ee6e"} Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.768843 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.891524 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-utilities\") pod \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.892014 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqbq5\" (UniqueName: \"kubernetes.io/projected/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-kube-api-access-gqbq5\") pod \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.892053 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-catalog-content\") pod \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\" (UID: \"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce\") " Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.893031 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-utilities" (OuterVolumeSpecName: "utilities") pod "fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" (UID: "fb6dd66e-7ccf-417e-9ae8-f837eee4cdce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.898412 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-kube-api-access-gqbq5" (OuterVolumeSpecName: "kube-api-access-gqbq5") pod "fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" (UID: "fb6dd66e-7ccf-417e-9ae8-f837eee4cdce"). InnerVolumeSpecName "kube-api-access-gqbq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.944138 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" (UID: "fb6dd66e-7ccf-417e-9ae8-f837eee4cdce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.993531 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqbq5\" (UniqueName: \"kubernetes.io/projected/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-kube-api-access-gqbq5\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.993569 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:48 crc kubenswrapper[4796]: I1127 11:41:48.993579 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.725309 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-75kwt" event={"ID":"fb6dd66e-7ccf-417e-9ae8-f837eee4cdce","Type":"ContainerDied","Data":"7594686ea46f74a507e697443e0760396664c3e60960acb35216fed9de0e982d"} Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.725361 4796 scope.go:117] "RemoveContainer" containerID="7e759e73b228715e0cb9bfe35cc7ff831cafb0dcc2f69a7b17b5d14ae199ee6e" Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.725427 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-75kwt" Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.760872 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-75kwt"] Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.761540 4796 scope.go:117] "RemoveContainer" containerID="c6ca6694c403573d67ef123ce7f86f6aad2782355bc0d96e17a133a9dab7f010" Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.764968 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-75kwt"] Nov 27 11:41:49 crc kubenswrapper[4796]: I1127 11:41:49.785237 4796 scope.go:117] "RemoveContainer" containerID="33f0fb1cea4f41dbd8dff0f0fd8216566098853cc8343ea202a69c9cfe5cd765" Nov 27 11:41:50 crc kubenswrapper[4796]: I1127 11:41:50.609403 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9xxqp" Nov 27 11:41:51 crc kubenswrapper[4796]: I1127 11:41:51.582818 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" path="/var/lib/kubelet/pods/fb6dd66e-7ccf-417e-9ae8-f837eee4cdce/volumes" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.899008 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nwt9g"] Nov 27 11:41:52 crc kubenswrapper[4796]: E1127 11:41:52.899370 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="extract-utilities" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.899391 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="extract-utilities" Nov 27 11:41:52 crc kubenswrapper[4796]: E1127 11:41:52.899411 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="registry-server" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.899423 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="registry-server" Nov 27 11:41:52 crc kubenswrapper[4796]: E1127 11:41:52.899441 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="extract-content" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.899454 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="extract-content" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.899621 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb6dd66e-7ccf-417e-9ae8-f837eee4cdce" containerName="registry-server" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.900984 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:52 crc kubenswrapper[4796]: I1127 11:41:52.919956 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwt9g"] Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.044151 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-catalog-content\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.044416 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq2x2\" (UniqueName: \"kubernetes.io/projected/78130259-d5b2-493b-94c6-886a8484c5f5-kube-api-access-cq2x2\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.044458 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-utilities\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.145694 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-catalog-content\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.145799 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq2x2\" (UniqueName: \"kubernetes.io/projected/78130259-d5b2-493b-94c6-886a8484c5f5-kube-api-access-cq2x2\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.145822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-utilities\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.146333 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-utilities\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.146341 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-catalog-content\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.169074 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq2x2\" (UniqueName: \"kubernetes.io/projected/78130259-d5b2-493b-94c6-886a8484c5f5-kube-api-access-cq2x2\") pod \"redhat-operators-nwt9g\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.233193 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.463823 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwt9g"] Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.756069 4796 generic.go:334] "Generic (PLEG): container finished" podID="78130259-d5b2-493b-94c6-886a8484c5f5" containerID="2c3437ea8a6ce6bd1ab296f92ab28dc34a6e0b5329ee711e0472d34145f4d9e4" exitCode=0 Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.756114 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwt9g" event={"ID":"78130259-d5b2-493b-94c6-886a8484c5f5","Type":"ContainerDied","Data":"2c3437ea8a6ce6bd1ab296f92ab28dc34a6e0b5329ee711e0472d34145f4d9e4"} Nov 27 11:41:53 crc kubenswrapper[4796]: I1127 11:41:53.756442 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwt9g" event={"ID":"78130259-d5b2-493b-94c6-886a8484c5f5","Type":"ContainerStarted","Data":"3d7ea8e5ad6bd5c9afb9a16e0b4f7a3cd6efb91161ee931e1d965cfcb30cf978"} Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.667098 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zr4wp"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.671092 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.684845 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr4wp"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.769419 4796 generic.go:334] "Generic (PLEG): container finished" podID="78130259-d5b2-493b-94c6-886a8484c5f5" containerID="83388f8b2ec5b12262e56b5318ca5f758fabbf2bdeaae3c4b3f0aab155a6d0f3" exitCode=0 Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.769527 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwt9g" event={"ID":"78130259-d5b2-493b-94c6-886a8484c5f5","Type":"ContainerDied","Data":"83388f8b2ec5b12262e56b5318ca5f758fabbf2bdeaae3c4b3f0aab155a6d0f3"} Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.777974 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zhsg\" (UniqueName: \"kubernetes.io/projected/db0bde27-1355-455e-8eb7-72c1cb7740c3-kube-api-access-5zhsg\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.778077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-utilities\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.778126 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-catalog-content\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.879470 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zhsg\" (UniqueName: \"kubernetes.io/projected/db0bde27-1355-455e-8eb7-72c1cb7740c3-kube-api-access-5zhsg\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.879576 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-utilities\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.879608 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-catalog-content\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.880112 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-catalog-content\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.880219 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-utilities\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.904978 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zhsg\" (UniqueName: \"kubernetes.io/projected/db0bde27-1355-455e-8eb7-72c1cb7740c3-kube-api-access-5zhsg\") pod \"redhat-marketplace-zr4wp\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.925005 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zjbvh"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.925229 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zjbvh" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="registry-server" containerID="cri-o://03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a" gracePeriod=30 Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.940077 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-66p7v"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.940523 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-66p7v" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="registry-server" containerID="cri-o://7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b" gracePeriod=30 Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.949037 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lv29n"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.949593 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" podUID="097c089f-7300-463f-abc6-2f00735c18d9" containerName="marketplace-operator" containerID="cri-o://c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14" gracePeriod=30 Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.957733 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpct5"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.957965 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xpct5" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="registry-server" containerID="cri-o://8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3" gracePeriod=30 Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.967564 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr4wp"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.967946 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.988406 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pnvmc"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.989306 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.994547 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jxxzb"] Nov 27 11:41:55 crc kubenswrapper[4796]: I1127 11:41:55.994784 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jxxzb" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="registry-server" containerID="cri-o://740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b" gracePeriod=30 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.014893 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pnvmc"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.018881 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nwt9g"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.083503 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d2dj\" (UniqueName: \"kubernetes.io/projected/4a5aa833-5ce1-4847-bd66-45dcae6a2015-kube-api-access-7d2dj\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.083930 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4a5aa833-5ce1-4847-bd66-45dcae6a2015-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.083956 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a5aa833-5ce1-4847-bd66-45dcae6a2015-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.185500 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d2dj\" (UniqueName: \"kubernetes.io/projected/4a5aa833-5ce1-4847-bd66-45dcae6a2015-kube-api-access-7d2dj\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.185541 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4a5aa833-5ce1-4847-bd66-45dcae6a2015-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.185564 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a5aa833-5ce1-4847-bd66-45dcae6a2015-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.189063 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a5aa833-5ce1-4847-bd66-45dcae6a2015-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.193541 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4a5aa833-5ce1-4847-bd66-45dcae6a2015-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.203258 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d2dj\" (UniqueName: \"kubernetes.io/projected/4a5aa833-5ce1-4847-bd66-45dcae6a2015-kube-api-access-7d2dj\") pod \"marketplace-operator-79b997595-pnvmc\" (UID: \"4a5aa833-5ce1-4847-bd66-45dcae6a2015\") " pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.364503 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.388365 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr4wp"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.393696 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:41:56 crc kubenswrapper[4796]: W1127 11:41:56.421684 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb0bde27_1355_455e_8eb7_72c1cb7740c3.slice/crio-5165149fdfae2077f7ad7350769af9dc005e2cbbb6e30bf86630fb28bcd30812 WatchSource:0}: Error finding container 5165149fdfae2077f7ad7350769af9dc005e2cbbb6e30bf86630fb28bcd30812: Status 404 returned error can't find the container with id 5165149fdfae2077f7ad7350769af9dc005e2cbbb6e30bf86630fb28bcd30812 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.445582 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.446391 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.448839 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.489203 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-utilities\") pod \"5e88b470-90dd-4759-939d-65edfde8e6f6\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.489257 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt4x7\" (UniqueName: \"kubernetes.io/projected/5e88b470-90dd-4759-939d-65edfde8e6f6-kube-api-access-mt4x7\") pod \"5e88b470-90dd-4759-939d-65edfde8e6f6\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.489399 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-catalog-content\") pod \"5e88b470-90dd-4759-939d-65edfde8e6f6\" (UID: \"5e88b470-90dd-4759-939d-65edfde8e6f6\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.490747 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-utilities" (OuterVolumeSpecName: "utilities") pod "5e88b470-90dd-4759-939d-65edfde8e6f6" (UID: "5e88b470-90dd-4759-939d-65edfde8e6f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.493551 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e88b470-90dd-4759-939d-65edfde8e6f6-kube-api-access-mt4x7" (OuterVolumeSpecName: "kube-api-access-mt4x7") pod "5e88b470-90dd-4759-939d-65edfde8e6f6" (UID: "5e88b470-90dd-4759-939d-65edfde8e6f6"). InnerVolumeSpecName "kube-api-access-mt4x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.558122 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e88b470-90dd-4759-939d-65edfde8e6f6" (UID: "5e88b470-90dd-4759-939d-65edfde8e6f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.590198 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqhv6\" (UniqueName: \"kubernetes.io/projected/79f62242-42c1-459f-9828-00f28d5de84d-kube-api-access-jqhv6\") pod \"79f62242-42c1-459f-9828-00f28d5de84d\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.590309 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-trusted-ca\") pod \"097c089f-7300-463f-abc6-2f00735c18d9\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.590388 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5qxs\" (UniqueName: \"kubernetes.io/projected/667350f0-abe3-4fc2-8d32-f817daf72697-kube-api-access-n5qxs\") pod \"667350f0-abe3-4fc2-8d32-f817daf72697\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591137 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-catalog-content\") pod \"79f62242-42c1-459f-9828-00f28d5de84d\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591158 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-operator-metrics\") pod \"097c089f-7300-463f-abc6-2f00735c18d9\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591189 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-utilities\") pod \"79f62242-42c1-459f-9828-00f28d5de84d\" (UID: \"79f62242-42c1-459f-9828-00f28d5de84d\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591226 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-catalog-content\") pod \"667350f0-abe3-4fc2-8d32-f817daf72697\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591278 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-utilities\") pod \"667350f0-abe3-4fc2-8d32-f817daf72697\" (UID: \"667350f0-abe3-4fc2-8d32-f817daf72697\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591301 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6crk\" (UniqueName: \"kubernetes.io/projected/097c089f-7300-463f-abc6-2f00735c18d9-kube-api-access-m6crk\") pod \"097c089f-7300-463f-abc6-2f00735c18d9\" (UID: \"097c089f-7300-463f-abc6-2f00735c18d9\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591488 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591499 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e88b470-90dd-4759-939d-65edfde8e6f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.591507 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt4x7\" (UniqueName: \"kubernetes.io/projected/5e88b470-90dd-4759-939d-65edfde8e6f6-kube-api-access-mt4x7\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.592393 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "097c089f-7300-463f-abc6-2f00735c18d9" (UID: "097c089f-7300-463f-abc6-2f00735c18d9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.594201 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "097c089f-7300-463f-abc6-2f00735c18d9" (UID: "097c089f-7300-463f-abc6-2f00735c18d9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.594314 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-utilities" (OuterVolumeSpecName: "utilities") pod "79f62242-42c1-459f-9828-00f28d5de84d" (UID: "79f62242-42c1-459f-9828-00f28d5de84d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.594695 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-utilities" (OuterVolumeSpecName: "utilities") pod "667350f0-abe3-4fc2-8d32-f817daf72697" (UID: "667350f0-abe3-4fc2-8d32-f817daf72697"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.594779 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097c089f-7300-463f-abc6-2f00735c18d9-kube-api-access-m6crk" (OuterVolumeSpecName: "kube-api-access-m6crk") pod "097c089f-7300-463f-abc6-2f00735c18d9" (UID: "097c089f-7300-463f-abc6-2f00735c18d9"). InnerVolumeSpecName "kube-api-access-m6crk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.595045 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667350f0-abe3-4fc2-8d32-f817daf72697-kube-api-access-n5qxs" (OuterVolumeSpecName: "kube-api-access-n5qxs") pod "667350f0-abe3-4fc2-8d32-f817daf72697" (UID: "667350f0-abe3-4fc2-8d32-f817daf72697"). InnerVolumeSpecName "kube-api-access-n5qxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.596946 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f62242-42c1-459f-9828-00f28d5de84d-kube-api-access-jqhv6" (OuterVolumeSpecName: "kube-api-access-jqhv6") pod "79f62242-42c1-459f-9828-00f28d5de84d" (UID: "79f62242-42c1-459f-9828-00f28d5de84d"). InnerVolumeSpecName "kube-api-access-jqhv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.655598 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "667350f0-abe3-4fc2-8d32-f817daf72697" (UID: "667350f0-abe3-4fc2-8d32-f817daf72697"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.688588 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79f62242-42c1-459f-9828-00f28d5de84d" (UID: "79f62242-42c1-459f-9828-00f28d5de84d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.692906 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.692946 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6crk\" (UniqueName: \"kubernetes.io/projected/097c089f-7300-463f-abc6-2f00735c18d9-kube-api-access-m6crk\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.692963 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqhv6\" (UniqueName: \"kubernetes.io/projected/79f62242-42c1-459f-9828-00f28d5de84d-kube-api-access-jqhv6\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.692975 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.692987 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5qxs\" (UniqueName: \"kubernetes.io/projected/667350f0-abe3-4fc2-8d32-f817daf72697-kube-api-access-n5qxs\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.692998 4796 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/097c089f-7300-463f-abc6-2f00735c18d9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.693012 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.693022 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f62242-42c1-459f-9828-00f28d5de84d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.693032 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/667350f0-abe3-4fc2-8d32-f817daf72697-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.760700 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.776757 4796 generic.go:334] "Generic (PLEG): container finished" podID="db0bde27-1355-455e-8eb7-72c1cb7740c3" containerID="6f70533f591eaa288e20e65c0ed4e6d16a66bb52de762db093d31281e3c65379" exitCode=0 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.776826 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr4wp" event={"ID":"db0bde27-1355-455e-8eb7-72c1cb7740c3","Type":"ContainerDied","Data":"6f70533f591eaa288e20e65c0ed4e6d16a66bb52de762db093d31281e3c65379"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.778120 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr4wp" event={"ID":"db0bde27-1355-455e-8eb7-72c1cb7740c3","Type":"ContainerStarted","Data":"5165149fdfae2077f7ad7350769af9dc005e2cbbb6e30bf86630fb28bcd30812"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.787567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwt9g" event={"ID":"78130259-d5b2-493b-94c6-886a8484c5f5","Type":"ContainerStarted","Data":"4c311dd2080f1b86d69d773ad6bffa4ce86591e2cd68a900a43f15eebd08b980"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.787638 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nwt9g" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="registry-server" containerID="cri-o://4c311dd2080f1b86d69d773ad6bffa4ce86591e2cd68a900a43f15eebd08b980" gracePeriod=30 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.793678 4796 generic.go:334] "Generic (PLEG): container finished" podID="79f62242-42c1-459f-9828-00f28d5de84d" containerID="740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b" exitCode=0 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.793738 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxxzb" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.793745 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerDied","Data":"740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.794092 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxxzb" event={"ID":"79f62242-42c1-459f-9828-00f28d5de84d","Type":"ContainerDied","Data":"6037a24fbbf813ab07cf808ccfba4524366e67ecb9488778ea08812f82a4e5b4"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.794117 4796 scope.go:117] "RemoveContainer" containerID="740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.796778 4796 generic.go:334] "Generic (PLEG): container finished" podID="667350f0-abe3-4fc2-8d32-f817daf72697" containerID="7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b" exitCode=0 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.796837 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerDied","Data":"7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.796868 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66p7v" event={"ID":"667350f0-abe3-4fc2-8d32-f817daf72697","Type":"ContainerDied","Data":"6487b9c637ed73666a8a1f6b3004267d73e17c25bca08da5b4decfd180363857"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.796930 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66p7v" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.809341 4796 generic.go:334] "Generic (PLEG): container finished" podID="097c089f-7300-463f-abc6-2f00735c18d9" containerID="c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14" exitCode=0 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.809471 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" event={"ID":"097c089f-7300-463f-abc6-2f00735c18d9","Type":"ContainerDied","Data":"c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.809523 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" event={"ID":"097c089f-7300-463f-abc6-2f00735c18d9","Type":"ContainerDied","Data":"5c851084da1c87d135d804b42531741beb2b37caacfdaaf333d70b7750067679"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.809471 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-lv29n" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.811044 4796 scope.go:117] "RemoveContainer" containerID="a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.818567 4796 generic.go:334] "Generic (PLEG): container finished" podID="6b58a0da-df89-4f60-9152-1a56e3604878" containerID="8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3" exitCode=0 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.818668 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerDied","Data":"8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.818699 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xpct5" event={"ID":"6b58a0da-df89-4f60-9152-1a56e3604878","Type":"ContainerDied","Data":"6571be8aeff516067da7d9b297bdc9609cc45220ee822093323e59cdeba6b050"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.818784 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xpct5" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.825188 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nwt9g" podStartSLOduration=2.342432079 podStartE2EDuration="4.825120838s" podCreationTimestamp="2025-11-27 11:41:52 +0000 UTC" firstStartedPulling="2025-11-27 11:41:53.757279464 +0000 UTC m=+1031.275598382" lastFinishedPulling="2025-11-27 11:41:56.239968223 +0000 UTC m=+1033.758287141" observedRunningTime="2025-11-27 11:41:56.819248958 +0000 UTC m=+1034.337567876" watchObservedRunningTime="2025-11-27 11:41:56.825120838 +0000 UTC m=+1034.343439766" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.833504 4796 generic.go:334] "Generic (PLEG): container finished" podID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerID="03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a" exitCode=0 Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.833557 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerDied","Data":"03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.833594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjbvh" event={"ID":"5e88b470-90dd-4759-939d-65edfde8e6f6","Type":"ContainerDied","Data":"c2fd91f9452d3756a5d937cc52377667286165df8ebf030408998c1597ad0420"} Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.833635 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjbvh" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.849218 4796 scope.go:117] "RemoveContainer" containerID="c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.849671 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lv29n"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.854777 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-lv29n"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.862132 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pnvmc"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.874504 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jxxzb"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.878254 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jxxzb"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.881330 4796 scope.go:117] "RemoveContainer" containerID="740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b" Nov 27 11:41:56 crc kubenswrapper[4796]: W1127 11:41:56.883514 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a5aa833_5ce1_4847_bd66_45dcae6a2015.slice/crio-96d9e75c02552f648a2a32e0e92c17785e8d511a054298759a3951773a34df51 WatchSource:0}: Error finding container 96d9e75c02552f648a2a32e0e92c17785e8d511a054298759a3951773a34df51: Status 404 returned error can't find the container with id 96d9e75c02552f648a2a32e0e92c17785e8d511a054298759a3951773a34df51 Nov 27 11:41:56 crc kubenswrapper[4796]: E1127 11:41:56.883576 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b\": container with ID starting with 740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b not found: ID does not exist" containerID="740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.883611 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b"} err="failed to get container status \"740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b\": rpc error: code = NotFound desc = could not find container \"740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b\": container with ID starting with 740cf148ad754e42d616bb203d043434f45afbabda6c7801a70f070c61fbaa8b not found: ID does not exist" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.883654 4796 scope.go:117] "RemoveContainer" containerID="a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21" Nov 27 11:41:56 crc kubenswrapper[4796]: E1127 11:41:56.884565 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21\": container with ID starting with a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21 not found: ID does not exist" containerID="a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.884624 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21"} err="failed to get container status \"a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21\": rpc error: code = NotFound desc = could not find container \"a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21\": container with ID starting with a3f575d13bf91328f3aeb7a7ff536ea76f88956447d588f1233ed5ace557ad21 not found: ID does not exist" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.884656 4796 scope.go:117] "RemoveContainer" containerID="c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb" Nov 27 11:41:56 crc kubenswrapper[4796]: E1127 11:41:56.888721 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb\": container with ID starting with c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb not found: ID does not exist" containerID="c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.888774 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb"} err="failed to get container status \"c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb\": rpc error: code = NotFound desc = could not find container \"c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb\": container with ID starting with c0a9aa2aeb59a249f3dc0651ed8ae1063f7d2ff29818c53cedd59a2cfb9d3eeb not found: ID does not exist" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.888795 4796 scope.go:117] "RemoveContainer" containerID="7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.892714 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-66p7v"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.894359 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-utilities\") pod \"6b58a0da-df89-4f60-9152-1a56e3604878\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.894422 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5qmc\" (UniqueName: \"kubernetes.io/projected/6b58a0da-df89-4f60-9152-1a56e3604878-kube-api-access-n5qmc\") pod \"6b58a0da-df89-4f60-9152-1a56e3604878\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.894456 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-catalog-content\") pod \"6b58a0da-df89-4f60-9152-1a56e3604878\" (UID: \"6b58a0da-df89-4f60-9152-1a56e3604878\") " Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.896102 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-66p7v"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.907698 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-utilities" (OuterVolumeSpecName: "utilities") pod "6b58a0da-df89-4f60-9152-1a56e3604878" (UID: "6b58a0da-df89-4f60-9152-1a56e3604878"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.908134 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b58a0da-df89-4f60-9152-1a56e3604878-kube-api-access-n5qmc" (OuterVolumeSpecName: "kube-api-access-n5qmc") pod "6b58a0da-df89-4f60-9152-1a56e3604878" (UID: "6b58a0da-df89-4f60-9152-1a56e3604878"). InnerVolumeSpecName "kube-api-access-n5qmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.918118 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zjbvh"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.918169 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zjbvh"] Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.923050 4796 scope.go:117] "RemoveContainer" containerID="db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.933700 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b58a0da-df89-4f60-9152-1a56e3604878" (UID: "6b58a0da-df89-4f60-9152-1a56e3604878"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.995423 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.995456 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5qmc\" (UniqueName: \"kubernetes.io/projected/6b58a0da-df89-4f60-9152-1a56e3604878-kube-api-access-n5qmc\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:56 crc kubenswrapper[4796]: I1127 11:41:56.995465 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b58a0da-df89-4f60-9152-1a56e3604878-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.017755 4796 scope.go:117] "RemoveContainer" containerID="c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.042196 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.046816 4796 scope.go:117] "RemoveContainer" containerID="7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.053523 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b\": container with ID starting with 7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b not found: ID does not exist" containerID="7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.053569 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b"} err="failed to get container status \"7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b\": rpc error: code = NotFound desc = could not find container \"7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b\": container with ID starting with 7128744cca0abbd436315d47532afdb51151d52c3842cbab1d8b0c1affdf6a1b not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.053592 4796 scope.go:117] "RemoveContainer" containerID="db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.053918 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f\": container with ID starting with db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f not found: ID does not exist" containerID="db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.053960 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f"} err="failed to get container status \"db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f\": rpc error: code = NotFound desc = could not find container \"db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f\": container with ID starting with db878e5d15a2614975e1dbba5175eb1a9c97e8db2fe873ddca6ccb08a962d40f not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.053974 4796 scope.go:117] "RemoveContainer" containerID="c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.054175 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713\": container with ID starting with c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713 not found: ID does not exist" containerID="c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.054197 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713"} err="failed to get container status \"c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713\": rpc error: code = NotFound desc = could not find container \"c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713\": container with ID starting with c69b849f3e73d82f14f8fe8f7cf77809b63afbd04b0e41ef83b69b68b3c3e713 not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.054223 4796 scope.go:117] "RemoveContainer" containerID="c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.070242 4796 scope.go:117] "RemoveContainer" containerID="c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.070767 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14\": container with ID starting with c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14 not found: ID does not exist" containerID="c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.070792 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14"} err="failed to get container status \"c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14\": rpc error: code = NotFound desc = could not find container \"c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14\": container with ID starting with c781045d02eff5a6ea2f6642cc10d99b6bdd44d38715244cc325a12dd9500a14 not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.070810 4796 scope.go:117] "RemoveContainer" containerID="8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.083584 4796 scope.go:117] "RemoveContainer" containerID="6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.098829 4796 scope.go:117] "RemoveContainer" containerID="e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.111822 4796 scope.go:117] "RemoveContainer" containerID="8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.112218 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3\": container with ID starting with 8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3 not found: ID does not exist" containerID="8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.112301 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3"} err="failed to get container status \"8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3\": rpc error: code = NotFound desc = could not find container \"8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3\": container with ID starting with 8302036a6699851f967863b66ae1ed8812c313c4c8d01ad465fffc3d6c3bbfd3 not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.112324 4796 scope.go:117] "RemoveContainer" containerID="6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.112669 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7\": container with ID starting with 6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7 not found: ID does not exist" containerID="6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.112702 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7"} err="failed to get container status \"6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7\": rpc error: code = NotFound desc = could not find container \"6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7\": container with ID starting with 6e644311b387aa890e724f9dd5a08d4c5b1c2a1fad595bf2e20ba8293147b4c7 not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.112728 4796 scope.go:117] "RemoveContainer" containerID="e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.113091 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf\": container with ID starting with e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf not found: ID does not exist" containerID="e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.113115 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf"} err="failed to get container status \"e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf\": rpc error: code = NotFound desc = could not find container \"e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf\": container with ID starting with e39a36fd6a038081af1d1ae01986ada7f6924b1094dc6e05321660310908cadf not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.113130 4796 scope.go:117] "RemoveContainer" containerID="03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.182034 4796 scope.go:117] "RemoveContainer" containerID="2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.200785 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpct5"] Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.200873 4796 scope.go:117] "RemoveContainer" containerID="1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.201387 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zhsg\" (UniqueName: \"kubernetes.io/projected/db0bde27-1355-455e-8eb7-72c1cb7740c3-kube-api-access-5zhsg\") pod \"db0bde27-1355-455e-8eb7-72c1cb7740c3\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.201416 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-catalog-content\") pod \"db0bde27-1355-455e-8eb7-72c1cb7740c3\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.201529 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-utilities\") pod \"db0bde27-1355-455e-8eb7-72c1cb7740c3\" (UID: \"db0bde27-1355-455e-8eb7-72c1cb7740c3\") " Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.201824 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xpct5"] Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.202443 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db0bde27-1355-455e-8eb7-72c1cb7740c3" (UID: "db0bde27-1355-455e-8eb7-72c1cb7740c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.205443 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-utilities" (OuterVolumeSpecName: "utilities") pod "db0bde27-1355-455e-8eb7-72c1cb7740c3" (UID: "db0bde27-1355-455e-8eb7-72c1cb7740c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.206686 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0bde27-1355-455e-8eb7-72c1cb7740c3-kube-api-access-5zhsg" (OuterVolumeSpecName: "kube-api-access-5zhsg") pod "db0bde27-1355-455e-8eb7-72c1cb7740c3" (UID: "db0bde27-1355-455e-8eb7-72c1cb7740c3"). InnerVolumeSpecName "kube-api-access-5zhsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.222583 4796 scope.go:117] "RemoveContainer" containerID="03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.223621 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a\": container with ID starting with 03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a not found: ID does not exist" containerID="03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.223659 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a"} err="failed to get container status \"03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a\": rpc error: code = NotFound desc = could not find container \"03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a\": container with ID starting with 03bb041de45d77dac03c48bad1cc052a591b8260429f03fa6e9821acb275094a not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.223695 4796 scope.go:117] "RemoveContainer" containerID="2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.223978 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35\": container with ID starting with 2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35 not found: ID does not exist" containerID="2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.224017 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35"} err="failed to get container status \"2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35\": rpc error: code = NotFound desc = could not find container \"2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35\": container with ID starting with 2fb0a3af0665754f1d8289e6260d440f9ca693ee4ebea67bf33b9b7869a55d35 not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.224043 4796 scope.go:117] "RemoveContainer" containerID="1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b" Nov 27 11:41:57 crc kubenswrapper[4796]: E1127 11:41:57.224418 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b\": container with ID starting with 1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b not found: ID does not exist" containerID="1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.224445 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b"} err="failed to get container status \"1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b\": rpc error: code = NotFound desc = could not find container \"1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b\": container with ID starting with 1cb6410e5e27b021fb9db0d22df38bf2275eceaa60afc53b4abfe5b647ccfe1b not found: ID does not exist" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.302508 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.302546 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zhsg\" (UniqueName: \"kubernetes.io/projected/db0bde27-1355-455e-8eb7-72c1cb7740c3-kube-api-access-5zhsg\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.302555 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db0bde27-1355-455e-8eb7-72c1cb7740c3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.581602 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097c089f-7300-463f-abc6-2f00735c18d9" path="/var/lib/kubelet/pods/097c089f-7300-463f-abc6-2f00735c18d9/volumes" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.582639 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" path="/var/lib/kubelet/pods/5e88b470-90dd-4759-939d-65edfde8e6f6/volumes" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.583833 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" path="/var/lib/kubelet/pods/667350f0-abe3-4fc2-8d32-f817daf72697/volumes" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.585489 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" path="/var/lib/kubelet/pods/6b58a0da-df89-4f60-9152-1a56e3604878/volumes" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.586090 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f62242-42c1-459f-9828-00f28d5de84d" path="/var/lib/kubelet/pods/79f62242-42c1-459f-9828-00f28d5de84d/volumes" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.843570 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" event={"ID":"4a5aa833-5ce1-4847-bd66-45dcae6a2015","Type":"ContainerStarted","Data":"0948286b8bf436ff6774fa81de2813de0a8ec261ef5fb8298818e52f76f31db9"} Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.843616 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" event={"ID":"4a5aa833-5ce1-4847-bd66-45dcae6a2015","Type":"ContainerStarted","Data":"96d9e75c02552f648a2a32e0e92c17785e8d511a054298759a3951773a34df51"} Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.844065 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.856436 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.868536 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-pnvmc" podStartSLOduration=2.868513975 podStartE2EDuration="2.868513975s" podCreationTimestamp="2025-11-27 11:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:41:57.862713486 +0000 UTC m=+1035.381032414" watchObservedRunningTime="2025-11-27 11:41:57.868513975 +0000 UTC m=+1035.386832893" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.872848 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr4wp" event={"ID":"db0bde27-1355-455e-8eb7-72c1cb7740c3","Type":"ContainerDied","Data":"5165149fdfae2077f7ad7350769af9dc005e2cbbb6e30bf86630fb28bcd30812"} Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.872893 4796 scope.go:117] "RemoveContainer" containerID="6f70533f591eaa288e20e65c0ed4e6d16a66bb52de762db093d31281e3c65379" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.872854 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr4wp" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.876561 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nwt9g_78130259-d5b2-493b-94c6-886a8484c5f5/registry-server/0.log" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.877657 4796 generic.go:334] "Generic (PLEG): container finished" podID="78130259-d5b2-493b-94c6-886a8484c5f5" containerID="4c311dd2080f1b86d69d773ad6bffa4ce86591e2cd68a900a43f15eebd08b980" exitCode=1 Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.877689 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwt9g" event={"ID":"78130259-d5b2-493b-94c6-886a8484c5f5","Type":"ContainerDied","Data":"4c311dd2080f1b86d69d773ad6bffa4ce86591e2cd68a900a43f15eebd08b980"} Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.890464 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nwt9g_78130259-d5b2-493b-94c6-886a8484c5f5/registry-server/0.log" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.891730 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.922309 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr4wp"] Nov 27 11:41:57 crc kubenswrapper[4796]: I1127 11:41:57.929085 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr4wp"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.014339 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-catalog-content\") pod \"78130259-d5b2-493b-94c6-886a8484c5f5\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.014417 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq2x2\" (UniqueName: \"kubernetes.io/projected/78130259-d5b2-493b-94c6-886a8484c5f5-kube-api-access-cq2x2\") pod \"78130259-d5b2-493b-94c6-886a8484c5f5\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.014504 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-utilities\") pod \"78130259-d5b2-493b-94c6-886a8484c5f5\" (UID: \"78130259-d5b2-493b-94c6-886a8484c5f5\") " Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.015368 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-utilities" (OuterVolumeSpecName: "utilities") pod "78130259-d5b2-493b-94c6-886a8484c5f5" (UID: "78130259-d5b2-493b-94c6-886a8484c5f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.018374 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78130259-d5b2-493b-94c6-886a8484c5f5-kube-api-access-cq2x2" (OuterVolumeSpecName: "kube-api-access-cq2x2") pod "78130259-d5b2-493b-94c6-886a8484c5f5" (UID: "78130259-d5b2-493b-94c6-886a8484c5f5"). InnerVolumeSpecName "kube-api-access-cq2x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065414 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lz6dl"] Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065620 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065632 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065644 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065651 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065663 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065670 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065683 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065689 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065697 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0bde27-1355-455e-8eb7-72c1cb7740c3" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065702 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0bde27-1355-455e-8eb7-72c1cb7740c3" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065709 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097c089f-7300-463f-abc6-2f00735c18d9" containerName="marketplace-operator" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065715 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="097c089f-7300-463f-abc6-2f00735c18d9" containerName="marketplace-operator" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065723 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065729 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065739 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065746 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065754 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065762 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065772 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065779 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065787 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065795 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065807 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065814 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065822 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065828 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="extract-content" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065836 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065841 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065849 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065855 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065864 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065870 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: E1127 11:41:58.065881 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065888 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065977 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065990 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0bde27-1355-455e-8eb7-72c1cb7740c3" containerName="extract-utilities" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.065997 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b58a0da-df89-4f60-9152-1a56e3604878" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.066006 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f62242-42c1-459f-9828-00f28d5de84d" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.066015 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="097c089f-7300-463f-abc6-2f00735c18d9" containerName="marketplace-operator" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.066020 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e88b470-90dd-4759-939d-65edfde8e6f6" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.066030 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="667350f0-abe3-4fc2-8d32-f817daf72697" containerName="registry-server" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.067319 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.069783 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.077964 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lz6dl"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.116534 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq2x2\" (UniqueName: \"kubernetes.io/projected/78130259-d5b2-493b-94c6-886a8484c5f5-kube-api-access-cq2x2\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.116810 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.127607 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78130259-d5b2-493b-94c6-886a8484c5f5" (UID: "78130259-d5b2-493b-94c6-886a8484c5f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.218313 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dfmg\" (UniqueName: \"kubernetes.io/projected/90010dd3-da45-452c-bbf0-d82893af6d97-kube-api-access-2dfmg\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.218395 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90010dd3-da45-452c-bbf0-d82893af6d97-catalog-content\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.218451 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90010dd3-da45-452c-bbf0-d82893af6d97-utilities\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.218623 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78130259-d5b2-493b-94c6-886a8484c5f5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.320340 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dfmg\" (UniqueName: \"kubernetes.io/projected/90010dd3-da45-452c-bbf0-d82893af6d97-kube-api-access-2dfmg\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.320411 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90010dd3-da45-452c-bbf0-d82893af6d97-catalog-content\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.320450 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90010dd3-da45-452c-bbf0-d82893af6d97-utilities\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.321375 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90010dd3-da45-452c-bbf0-d82893af6d97-catalog-content\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.321445 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90010dd3-da45-452c-bbf0-d82893af6d97-utilities\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.353095 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dfmg\" (UniqueName: \"kubernetes.io/projected/90010dd3-da45-452c-bbf0-d82893af6d97-kube-api-access-2dfmg\") pod \"redhat-marketplace-lz6dl\" (UID: \"90010dd3-da45-452c-bbf0-d82893af6d97\") " pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.384618 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.468398 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9kk7f"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.470600 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.499045 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kk7f"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.605954 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lz6dl"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.625766 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-catalog-content\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.625834 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-utilities\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.626349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mrfd\" (UniqueName: \"kubernetes.io/projected/9a794975-9d80-47c9-bf8b-b26b71afa054-kube-api-access-4mrfd\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.727237 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-catalog-content\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.727343 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-utilities\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.727401 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mrfd\" (UniqueName: \"kubernetes.io/projected/9a794975-9d80-47c9-bf8b-b26b71afa054-kube-api-access-4mrfd\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.727847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-utilities\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.728094 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-catalog-content\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.746492 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mrfd\" (UniqueName: \"kubernetes.io/projected/9a794975-9d80-47c9-bf8b-b26b71afa054-kube-api-access-4mrfd\") pod \"redhat-marketplace-9kk7f\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.816260 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.885241 4796 generic.go:334] "Generic (PLEG): container finished" podID="90010dd3-da45-452c-bbf0-d82893af6d97" containerID="71b8f4f0aa48687f1549ecf5a30737382312c5a59bd6bbcc4bf98f952f82fe2e" exitCode=0 Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.886016 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lz6dl" event={"ID":"90010dd3-da45-452c-bbf0-d82893af6d97","Type":"ContainerDied","Data":"71b8f4f0aa48687f1549ecf5a30737382312c5a59bd6bbcc4bf98f952f82fe2e"} Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.886043 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lz6dl" event={"ID":"90010dd3-da45-452c-bbf0-d82893af6d97","Type":"ContainerStarted","Data":"b0646d5eae27c567ca1b926d4cd1796febd416164a61a7e3d32fa2a4807f7697"} Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.901570 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nwt9g_78130259-d5b2-493b-94c6-886a8484c5f5/registry-server/0.log" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.906398 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwt9g" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.907906 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwt9g" event={"ID":"78130259-d5b2-493b-94c6-886a8484c5f5","Type":"ContainerDied","Data":"3d7ea8e5ad6bd5c9afb9a16e0b4f7a3cd6efb91161ee931e1d965cfcb30cf978"} Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.907967 4796 scope.go:117] "RemoveContainer" containerID="4c311dd2080f1b86d69d773ad6bffa4ce86591e2cd68a900a43f15eebd08b980" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.952322 4796 scope.go:117] "RemoveContainer" containerID="83388f8b2ec5b12262e56b5318ca5f758fabbf2bdeaae3c4b3f0aab155a6d0f3" Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.963237 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nwt9g"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.967938 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nwt9g"] Nov 27 11:41:58 crc kubenswrapper[4796]: I1127 11:41:58.987972 4796 scope.go:117] "RemoveContainer" containerID="2c3437ea8a6ce6bd1ab296f92ab28dc34a6e0b5329ee711e0472d34145f4d9e4" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.071076 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mdp44"] Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.073402 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.076554 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.081670 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mdp44"] Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.233803 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4dqc\" (UniqueName: \"kubernetes.io/projected/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-kube-api-access-g4dqc\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.233889 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-catalog-content\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.233998 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-utilities\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: W1127 11:41:59.238162 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a794975_9d80_47c9_bf8b_b26b71afa054.slice/crio-3d4c1c7843ae83162a9b2f52c7c86c8e3040e870e80dbdb433ae33240d453dc0 WatchSource:0}: Error finding container 3d4c1c7843ae83162a9b2f52c7c86c8e3040e870e80dbdb433ae33240d453dc0: Status 404 returned error can't find the container with id 3d4c1c7843ae83162a9b2f52c7c86c8e3040e870e80dbdb433ae33240d453dc0 Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.238333 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kk7f"] Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.335214 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-catalog-content\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.335334 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-utilities\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.335392 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4dqc\" (UniqueName: \"kubernetes.io/projected/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-kube-api-access-g4dqc\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.335892 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-utilities\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.338240 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-catalog-content\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.362848 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4dqc\" (UniqueName: \"kubernetes.io/projected/8e43ee0f-5349-43b2-8d0c-d3f41ac9766a-kube-api-access-g4dqc\") pod \"certified-operators-mdp44\" (UID: \"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a\") " pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.397112 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.470526 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mqb6q"] Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.472006 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.491255 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqb6q"] Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.576650 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78130259-d5b2-493b-94c6-886a8484c5f5" path="/var/lib/kubelet/pods/78130259-d5b2-493b-94c6-886a8484c5f5/volumes" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.577687 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db0bde27-1355-455e-8eb7-72c1cb7740c3" path="/var/lib/kubelet/pods/db0bde27-1355-455e-8eb7-72c1cb7740c3/volumes" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.614353 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mdp44"] Nov 27 11:41:59 crc kubenswrapper[4796]: W1127 11:41:59.630386 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e43ee0f_5349_43b2_8d0c_d3f41ac9766a.slice/crio-274bb298088d5ad8cf0be5ffb9b23180030a17d23e43734af9af6134e65db90e WatchSource:0}: Error finding container 274bb298088d5ad8cf0be5ffb9b23180030a17d23e43734af9af6134e65db90e: Status 404 returned error can't find the container with id 274bb298088d5ad8cf0be5ffb9b23180030a17d23e43734af9af6134e65db90e Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.638011 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-catalog-content\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.638131 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-utilities\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.638208 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkttz\" (UniqueName: \"kubernetes.io/projected/cb789a35-ba63-4416-b70b-305b17fe7385-kube-api-access-rkttz\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.740153 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-catalog-content\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.740490 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-catalog-content\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.740731 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-utilities\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.740774 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkttz\" (UniqueName: \"kubernetes.io/projected/cb789a35-ba63-4416-b70b-305b17fe7385-kube-api-access-rkttz\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.741349 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-utilities\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.758469 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkttz\" (UniqueName: \"kubernetes.io/projected/cb789a35-ba63-4416-b70b-305b17fe7385-kube-api-access-rkttz\") pod \"certified-operators-mqb6q\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.801305 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.913326 4796 generic.go:334] "Generic (PLEG): container finished" podID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerID="6a730be87adb953babe47526a1b56418efc357179dca56013467063538d40af5" exitCode=0 Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.913416 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kk7f" event={"ID":"9a794975-9d80-47c9-bf8b-b26b71afa054","Type":"ContainerDied","Data":"6a730be87adb953babe47526a1b56418efc357179dca56013467063538d40af5"} Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.913822 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kk7f" event={"ID":"9a794975-9d80-47c9-bf8b-b26b71afa054","Type":"ContainerStarted","Data":"3d4c1c7843ae83162a9b2f52c7c86c8e3040e870e80dbdb433ae33240d453dc0"} Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.920742 4796 generic.go:334] "Generic (PLEG): container finished" podID="8e43ee0f-5349-43b2-8d0c-d3f41ac9766a" containerID="aaca28fd5687c0ef142d06fbe31f736893dbe1ee1e0886e954d602cf5a6d8aef" exitCode=0 Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.920812 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdp44" event={"ID":"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a","Type":"ContainerDied","Data":"aaca28fd5687c0ef142d06fbe31f736893dbe1ee1e0886e954d602cf5a6d8aef"} Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.920838 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdp44" event={"ID":"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a","Type":"ContainerStarted","Data":"274bb298088d5ad8cf0be5ffb9b23180030a17d23e43734af9af6134e65db90e"} Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.924843 4796 generic.go:334] "Generic (PLEG): container finished" podID="90010dd3-da45-452c-bbf0-d82893af6d97" containerID="f1bd503a72b31288dcafe9c56a0484ea0c6fa548007419c53ca578fbcebcaee9" exitCode=0 Nov 27 11:41:59 crc kubenswrapper[4796]: I1127 11:41:59.925011 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lz6dl" event={"ID":"90010dd3-da45-452c-bbf0-d82893af6d97","Type":"ContainerDied","Data":"f1bd503a72b31288dcafe9c56a0484ea0c6fa548007419c53ca578fbcebcaee9"} Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.024464 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqb6q"] Nov 27 11:42:00 crc kubenswrapper[4796]: W1127 11:42:00.033459 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb789a35_ba63_4416_b70b_305b17fe7385.slice/crio-40ea4e004a1d6d738870bf1b47a1e366778be7101bbc416fbdafc5dbb5c32c2b WatchSource:0}: Error finding container 40ea4e004a1d6d738870bf1b47a1e366778be7101bbc416fbdafc5dbb5c32c2b: Status 404 returned error can't find the container with id 40ea4e004a1d6d738870bf1b47a1e366778be7101bbc416fbdafc5dbb5c32c2b Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.872515 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f2pd5"] Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.874416 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.878592 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.885663 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2pd5"] Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.933574 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lz6dl" event={"ID":"90010dd3-da45-452c-bbf0-d82893af6d97","Type":"ContainerStarted","Data":"c8a5c06ce233bb1e9dbe1d098ffe55f2e7b0afc8104d696abbd8de05ce7ac345"} Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.934866 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb789a35-ba63-4416-b70b-305b17fe7385" containerID="ac41476e8ce136858e8935c2b1b500d1872ad09861f398ef4e9ab5fe39a20d04" exitCode=0 Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.934903 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerDied","Data":"ac41476e8ce136858e8935c2b1b500d1872ad09861f398ef4e9ab5fe39a20d04"} Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.934925 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerStarted","Data":"40ea4e004a1d6d738870bf1b47a1e366778be7101bbc416fbdafc5dbb5c32c2b"} Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.954864 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lz6dl" podStartSLOduration=1.463799697 podStartE2EDuration="2.954848776s" podCreationTimestamp="2025-11-27 11:41:58 +0000 UTC" firstStartedPulling="2025-11-27 11:41:58.901869517 +0000 UTC m=+1036.420188435" lastFinishedPulling="2025-11-27 11:42:00.392918596 +0000 UTC m=+1037.911237514" observedRunningTime="2025-11-27 11:42:00.950320581 +0000 UTC m=+1038.468639499" watchObservedRunningTime="2025-11-27 11:42:00.954848776 +0000 UTC m=+1038.473167694" Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.955109 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d4dfb4-bdf2-459f-8e07-5cd029849c42-utilities\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.955243 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d4dfb4-bdf2-459f-8e07-5cd029849c42-catalog-content\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:00 crc kubenswrapper[4796]: I1127 11:42:00.955340 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgd4\" (UniqueName: \"kubernetes.io/projected/54d4dfb4-bdf2-459f-8e07-5cd029849c42-kube-api-access-2xgd4\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.056391 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d4dfb4-bdf2-459f-8e07-5cd029849c42-utilities\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.056507 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d4dfb4-bdf2-459f-8e07-5cd029849c42-catalog-content\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.056568 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgd4\" (UniqueName: \"kubernetes.io/projected/54d4dfb4-bdf2-459f-8e07-5cd029849c42-kube-api-access-2xgd4\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.057016 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d4dfb4-bdf2-459f-8e07-5cd029849c42-utilities\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.057287 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d4dfb4-bdf2-459f-8e07-5cd029849c42-catalog-content\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.080305 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgd4\" (UniqueName: \"kubernetes.io/projected/54d4dfb4-bdf2-459f-8e07-5cd029849c42-kube-api-access-2xgd4\") pod \"community-operators-f2pd5\" (UID: \"54d4dfb4-bdf2-459f-8e07-5cd029849c42\") " pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.192768 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.459585 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2pd5"] Nov 27 11:42:01 crc kubenswrapper[4796]: W1127 11:42:01.466600 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54d4dfb4_bdf2_459f_8e07_5cd029849c42.slice/crio-99a045e0cb08f641a76061651325836b2582c1204ba2388fef322da870ff5553 WatchSource:0}: Error finding container 99a045e0cb08f641a76061651325836b2582c1204ba2388fef322da870ff5553: Status 404 returned error can't find the container with id 99a045e0cb08f641a76061651325836b2582c1204ba2388fef322da870ff5553 Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.867044 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b5fl4"] Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.876477 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.880792 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.880870 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.886104 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.894116 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5fl4"] Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.941576 4796 generic.go:334] "Generic (PLEG): container finished" podID="54d4dfb4-bdf2-459f-8e07-5cd029849c42" containerID="cb5b225b5014d85b1cac8de2399aeb1f64c4d1f764e44d467c1da3a5329c6959" exitCode=0 Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.941648 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2pd5" event={"ID":"54d4dfb4-bdf2-459f-8e07-5cd029849c42","Type":"ContainerDied","Data":"cb5b225b5014d85b1cac8de2399aeb1f64c4d1f764e44d467c1da3a5329c6959"} Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.941674 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2pd5" event={"ID":"54d4dfb4-bdf2-459f-8e07-5cd029849c42","Type":"ContainerStarted","Data":"99a045e0cb08f641a76061651325836b2582c1204ba2388fef322da870ff5553"} Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.943893 4796 generic.go:334] "Generic (PLEG): container finished" podID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerID="d32cbe6a4745c9700b5ece7fd9c472d7c27fbca27963b90f8029d6da44345fe6" exitCode=0 Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.943936 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kk7f" event={"ID":"9a794975-9d80-47c9-bf8b-b26b71afa054","Type":"ContainerDied","Data":"d32cbe6a4745c9700b5ece7fd9c472d7c27fbca27963b90f8029d6da44345fe6"} Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.951487 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdp44" event={"ID":"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a","Type":"ContainerStarted","Data":"8cedea12801320a0622b859f1914433e34db26a1df9bf784f63f40fc06381f29"} Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.966521 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-catalog-content\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.966570 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9mt\" (UniqueName: \"kubernetes.io/projected/e9abb298-96b8-41ea-b357-534bc4947657-kube-api-access-xx9mt\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:01 crc kubenswrapper[4796]: I1127 11:42:01.966605 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-utilities\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.067598 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-utilities\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.067705 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-catalog-content\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.067784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9mt\" (UniqueName: \"kubernetes.io/projected/e9abb298-96b8-41ea-b357-534bc4947657-kube-api-access-xx9mt\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.068210 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-utilities\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.068540 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-catalog-content\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.087610 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9mt\" (UniqueName: \"kubernetes.io/projected/e9abb298-96b8-41ea-b357-534bc4947657-kube-api-access-xx9mt\") pod \"redhat-operators-b5fl4\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.211937 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.601642 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5fl4"] Nov 27 11:42:02 crc kubenswrapper[4796]: W1127 11:42:02.607805 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9abb298_96b8_41ea_b357_534bc4947657.slice/crio-43133b8766e6acd7c61c39155f81c72a67e2272acabc9b6de6e319c520178bb1 WatchSource:0}: Error finding container 43133b8766e6acd7c61c39155f81c72a67e2272acabc9b6de6e319c520178bb1: Status 404 returned error can't find the container with id 43133b8766e6acd7c61c39155f81c72a67e2272acabc9b6de6e319c520178bb1 Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.958739 4796 generic.go:334] "Generic (PLEG): container finished" podID="8e43ee0f-5349-43b2-8d0c-d3f41ac9766a" containerID="8cedea12801320a0622b859f1914433e34db26a1df9bf784f63f40fc06381f29" exitCode=0 Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.958835 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdp44" event={"ID":"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a","Type":"ContainerDied","Data":"8cedea12801320a0622b859f1914433e34db26a1df9bf784f63f40fc06381f29"} Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.960772 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2pd5" event={"ID":"54d4dfb4-bdf2-459f-8e07-5cd029849c42","Type":"ContainerStarted","Data":"379e76dcbfa04be9979d07f4bb9189be82966298fcca31eb8802e68c0f68cb31"} Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.966050 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9abb298-96b8-41ea-b357-534bc4947657" containerID="0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17" exitCode=0 Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.966180 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerDied","Data":"0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17"} Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.966219 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerStarted","Data":"43133b8766e6acd7c61c39155f81c72a67e2272acabc9b6de6e319c520178bb1"} Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.968396 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerStarted","Data":"298a3753c34443e1af8e6f79911b277f52527e347f60bc3ef908fe16ec6bd01d"} Nov 27 11:42:02 crc kubenswrapper[4796]: I1127 11:42:02.971721 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kk7f" event={"ID":"9a794975-9d80-47c9-bf8b-b26b71afa054","Type":"ContainerStarted","Data":"86cc7ed9c2b0af55b501ee1c2d6d2632e6602eae24c8257e22f7fe4009eed7d1"} Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.014613 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9kk7f" podStartSLOduration=2.096126443 podStartE2EDuration="5.014596449s" podCreationTimestamp="2025-11-27 11:41:58 +0000 UTC" firstStartedPulling="2025-11-27 11:41:59.915999863 +0000 UTC m=+1037.434318781" lastFinishedPulling="2025-11-27 11:42:02.834469869 +0000 UTC m=+1040.352788787" observedRunningTime="2025-11-27 11:42:03.01098137 +0000 UTC m=+1040.529300318" watchObservedRunningTime="2025-11-27 11:42:03.014596449 +0000 UTC m=+1040.532915367" Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.979562 4796 generic.go:334] "Generic (PLEG): container finished" podID="54d4dfb4-bdf2-459f-8e07-5cd029849c42" containerID="379e76dcbfa04be9979d07f4bb9189be82966298fcca31eb8802e68c0f68cb31" exitCode=0 Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.979626 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2pd5" event={"ID":"54d4dfb4-bdf2-459f-8e07-5cd029849c42","Type":"ContainerDied","Data":"379e76dcbfa04be9979d07f4bb9189be82966298fcca31eb8802e68c0f68cb31"} Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.982371 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerStarted","Data":"dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab"} Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.988164 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb789a35-ba63-4416-b70b-305b17fe7385" containerID="298a3753c34443e1af8e6f79911b277f52527e347f60bc3ef908fe16ec6bd01d" exitCode=0 Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.988254 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerDied","Data":"298a3753c34443e1af8e6f79911b277f52527e347f60bc3ef908fe16ec6bd01d"} Nov 27 11:42:03 crc kubenswrapper[4796]: I1127 11:42:03.991812 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdp44" event={"ID":"8e43ee0f-5349-43b2-8d0c-d3f41ac9766a","Type":"ContainerStarted","Data":"a43080ca22d97c1676b3730d34940f440727479d2308e849bc379feec6dfce1b"} Nov 27 11:42:04 crc kubenswrapper[4796]: I1127 11:42:04.028628 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mdp44" podStartSLOduration=1.563647488 podStartE2EDuration="5.028599522s" podCreationTimestamp="2025-11-27 11:41:59 +0000 UTC" firstStartedPulling="2025-11-27 11:41:59.922345116 +0000 UTC m=+1037.440664034" lastFinishedPulling="2025-11-27 11:42:03.38729715 +0000 UTC m=+1040.905616068" observedRunningTime="2025-11-27 11:42:04.024115179 +0000 UTC m=+1041.542434097" watchObservedRunningTime="2025-11-27 11:42:04.028599522 +0000 UTC m=+1041.546918480" Nov 27 11:42:05 crc kubenswrapper[4796]: I1127 11:42:05.002199 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2pd5" event={"ID":"54d4dfb4-bdf2-459f-8e07-5cd029849c42","Type":"ContainerStarted","Data":"2193d32aa7810f8e07e726c927afdb97e8a5b04c577dd240df5fc33dfafd3feb"} Nov 27 11:42:05 crc kubenswrapper[4796]: I1127 11:42:05.004563 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9abb298-96b8-41ea-b357-534bc4947657" containerID="dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab" exitCode=0 Nov 27 11:42:05 crc kubenswrapper[4796]: I1127 11:42:05.004655 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerDied","Data":"dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab"} Nov 27 11:42:05 crc kubenswrapper[4796]: I1127 11:42:05.008329 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerStarted","Data":"ac9773d4deb566ad7d3576b7e67d3f5854b47b7eac48bf429ba3b8fca47bee02"} Nov 27 11:42:05 crc kubenswrapper[4796]: I1127 11:42:05.029608 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f2pd5" podStartSLOduration=2.499348648 podStartE2EDuration="5.029590149s" podCreationTimestamp="2025-11-27 11:42:00 +0000 UTC" firstStartedPulling="2025-11-27 11:42:01.943129794 +0000 UTC m=+1039.461448722" lastFinishedPulling="2025-11-27 11:42:04.473371305 +0000 UTC m=+1041.991690223" observedRunningTime="2025-11-27 11:42:05.026862694 +0000 UTC m=+1042.545181632" watchObservedRunningTime="2025-11-27 11:42:05.029590149 +0000 UTC m=+1042.547909067" Nov 27 11:42:05 crc kubenswrapper[4796]: I1127 11:42:05.053249 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mqb6q" podStartSLOduration=2.423705088 podStartE2EDuration="6.053231505s" podCreationTimestamp="2025-11-27 11:41:59 +0000 UTC" firstStartedPulling="2025-11-27 11:42:00.936919265 +0000 UTC m=+1038.455238183" lastFinishedPulling="2025-11-27 11:42:04.566445682 +0000 UTC m=+1042.084764600" observedRunningTime="2025-11-27 11:42:05.048698521 +0000 UTC m=+1042.567017449" watchObservedRunningTime="2025-11-27 11:42:05.053231505 +0000 UTC m=+1042.571550433" Nov 27 11:42:06 crc kubenswrapper[4796]: I1127 11:42:06.015860 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerStarted","Data":"ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023"} Nov 27 11:42:06 crc kubenswrapper[4796]: I1127 11:42:06.037334 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b5fl4" podStartSLOduration=2.446500391 podStartE2EDuration="5.037313038s" podCreationTimestamp="2025-11-27 11:42:01 +0000 UTC" firstStartedPulling="2025-11-27 11:42:02.967533272 +0000 UTC m=+1040.485852190" lastFinishedPulling="2025-11-27 11:42:05.558345919 +0000 UTC m=+1043.076664837" observedRunningTime="2025-11-27 11:42:06.032300962 +0000 UTC m=+1043.550619880" watchObservedRunningTime="2025-11-27 11:42:06.037313038 +0000 UTC m=+1043.555631966" Nov 27 11:42:08 crc kubenswrapper[4796]: I1127 11:42:08.385015 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:42:08 crc kubenswrapper[4796]: I1127 11:42:08.385502 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:42:08 crc kubenswrapper[4796]: I1127 11:42:08.436449 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:42:08 crc kubenswrapper[4796]: I1127 11:42:08.816895 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:42:08 crc kubenswrapper[4796]: I1127 11:42:08.816966 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:42:08 crc kubenswrapper[4796]: I1127 11:42:08.869530 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.090878 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lz6dl" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.116201 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.397970 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.398048 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.458159 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.802161 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.802211 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:42:09 crc kubenswrapper[4796]: I1127 11:42:09.866430 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:42:10 crc kubenswrapper[4796]: I1127 11:42:10.085191 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:42:10 crc kubenswrapper[4796]: I1127 11:42:10.103449 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mdp44" Nov 27 11:42:11 crc kubenswrapper[4796]: I1127 11:42:11.192962 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:11 crc kubenswrapper[4796]: I1127 11:42:11.193061 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:11 crc kubenswrapper[4796]: I1127 11:42:11.254073 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:11 crc kubenswrapper[4796]: I1127 11:42:11.266865 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kk7f"] Nov 27 11:42:11 crc kubenswrapper[4796]: I1127 11:42:11.267110 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9kk7f" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="registry-server" containerID="cri-o://86cc7ed9c2b0af55b501ee1c2d6d2632e6602eae24c8257e22f7fe4009eed7d1" gracePeriod=2 Nov 27 11:42:12 crc kubenswrapper[4796]: I1127 11:42:12.128906 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f2pd5" Nov 27 11:42:12 crc kubenswrapper[4796]: I1127 11:42:12.212518 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:12 crc kubenswrapper[4796]: I1127 11:42:12.212952 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:12 crc kubenswrapper[4796]: I1127 11:42:12.262250 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqb6q"] Nov 27 11:42:12 crc kubenswrapper[4796]: I1127 11:42:12.262544 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mqb6q" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="registry-server" containerID="cri-o://ac9773d4deb566ad7d3576b7e67d3f5854b47b7eac48bf429ba3b8fca47bee02" gracePeriod=2 Nov 27 11:42:12 crc kubenswrapper[4796]: I1127 11:42:12.278751 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.066844 4796 generic.go:334] "Generic (PLEG): container finished" podID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerID="86cc7ed9c2b0af55b501ee1c2d6d2632e6602eae24c8257e22f7fe4009eed7d1" exitCode=0 Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.066902 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kk7f" event={"ID":"9a794975-9d80-47c9-bf8b-b26b71afa054","Type":"ContainerDied","Data":"86cc7ed9c2b0af55b501ee1c2d6d2632e6602eae24c8257e22f7fe4009eed7d1"} Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.118028 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.289519 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.429162 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mrfd\" (UniqueName: \"kubernetes.io/projected/9a794975-9d80-47c9-bf8b-b26b71afa054-kube-api-access-4mrfd\") pod \"9a794975-9d80-47c9-bf8b-b26b71afa054\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.429346 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-utilities\") pod \"9a794975-9d80-47c9-bf8b-b26b71afa054\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.429418 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-catalog-content\") pod \"9a794975-9d80-47c9-bf8b-b26b71afa054\" (UID: \"9a794975-9d80-47c9-bf8b-b26b71afa054\") " Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.430620 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-utilities" (OuterVolumeSpecName: "utilities") pod "9a794975-9d80-47c9-bf8b-b26b71afa054" (UID: "9a794975-9d80-47c9-bf8b-b26b71afa054"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.436476 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a794975-9d80-47c9-bf8b-b26b71afa054-kube-api-access-4mrfd" (OuterVolumeSpecName: "kube-api-access-4mrfd") pod "9a794975-9d80-47c9-bf8b-b26b71afa054" (UID: "9a794975-9d80-47c9-bf8b-b26b71afa054"). InnerVolumeSpecName "kube-api-access-4mrfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.465500 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a794975-9d80-47c9-bf8b-b26b71afa054" (UID: "9a794975-9d80-47c9-bf8b-b26b71afa054"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.530841 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mrfd\" (UniqueName: \"kubernetes.io/projected/9a794975-9d80-47c9-bf8b-b26b71afa054-kube-api-access-4mrfd\") on node \"crc\" DevicePath \"\"" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.530878 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:42:13 crc kubenswrapper[4796]: I1127 11:42:13.530891 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a794975-9d80-47c9-bf8b-b26b71afa054-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.074897 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9kk7f" event={"ID":"9a794975-9d80-47c9-bf8b-b26b71afa054","Type":"ContainerDied","Data":"3d4c1c7843ae83162a9b2f52c7c86c8e3040e870e80dbdb433ae33240d453dc0"} Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.074901 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9kk7f" Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.074946 4796 scope.go:117] "RemoveContainer" containerID="86cc7ed9c2b0af55b501ee1c2d6d2632e6602eae24c8257e22f7fe4009eed7d1" Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.095954 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kk7f"] Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.099322 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9kk7f"] Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.115301 4796 scope.go:117] "RemoveContainer" containerID="d32cbe6a4745c9700b5ece7fd9c472d7c27fbca27963b90f8029d6da44345fe6" Nov 27 11:42:14 crc kubenswrapper[4796]: I1127 11:42:14.153740 4796 scope.go:117] "RemoveContainer" containerID="6a730be87adb953babe47526a1b56418efc357179dca56013467063538d40af5" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.083041 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb789a35-ba63-4416-b70b-305b17fe7385" containerID="ac9773d4deb566ad7d3576b7e67d3f5854b47b7eac48bf429ba3b8fca47bee02" exitCode=0 Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.083119 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerDied","Data":"ac9773d4deb566ad7d3576b7e67d3f5854b47b7eac48bf429ba3b8fca47bee02"} Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.576482 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" path="/var/lib/kubelet/pods/9a794975-9d80-47c9-bf8b-b26b71afa054/volumes" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.725148 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.864777 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-catalog-content\") pod \"cb789a35-ba63-4416-b70b-305b17fe7385\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.864976 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-utilities\") pod \"cb789a35-ba63-4416-b70b-305b17fe7385\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.865098 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkttz\" (UniqueName: \"kubernetes.io/projected/cb789a35-ba63-4416-b70b-305b17fe7385-kube-api-access-rkttz\") pod \"cb789a35-ba63-4416-b70b-305b17fe7385\" (UID: \"cb789a35-ba63-4416-b70b-305b17fe7385\") " Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.865943 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-utilities" (OuterVolumeSpecName: "utilities") pod "cb789a35-ba63-4416-b70b-305b17fe7385" (UID: "cb789a35-ba63-4416-b70b-305b17fe7385"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.872288 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb789a35-ba63-4416-b70b-305b17fe7385-kube-api-access-rkttz" (OuterVolumeSpecName: "kube-api-access-rkttz") pod "cb789a35-ba63-4416-b70b-305b17fe7385" (UID: "cb789a35-ba63-4416-b70b-305b17fe7385"). InnerVolumeSpecName "kube-api-access-rkttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.955174 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb789a35-ba63-4416-b70b-305b17fe7385" (UID: "cb789a35-ba63-4416-b70b-305b17fe7385"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.966919 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkttz\" (UniqueName: \"kubernetes.io/projected/cb789a35-ba63-4416-b70b-305b17fe7385-kube-api-access-rkttz\") on node \"crc\" DevicePath \"\"" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.966956 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:42:15 crc kubenswrapper[4796]: I1127 11:42:15.966966 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb789a35-ba63-4416-b70b-305b17fe7385-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.091906 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqb6q" event={"ID":"cb789a35-ba63-4416-b70b-305b17fe7385","Type":"ContainerDied","Data":"40ea4e004a1d6d738870bf1b47a1e366778be7101bbc416fbdafc5dbb5c32c2b"} Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.091951 4796 scope.go:117] "RemoveContainer" containerID="ac9773d4deb566ad7d3576b7e67d3f5854b47b7eac48bf429ba3b8fca47bee02" Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.092042 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqb6q" Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.112532 4796 scope.go:117] "RemoveContainer" containerID="298a3753c34443e1af8e6f79911b277f52527e347f60bc3ef908fe16ec6bd01d" Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.127965 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqb6q"] Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.132507 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mqb6q"] Nov 27 11:42:16 crc kubenswrapper[4796]: I1127 11:42:16.156339 4796 scope.go:117] "RemoveContainer" containerID="ac41476e8ce136858e8935c2b1b500d1872ad09861f398ef4e9ab5fe39a20d04" Nov 27 11:42:17 crc kubenswrapper[4796]: I1127 11:42:17.581123 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" path="/var/lib/kubelet/pods/cb789a35-ba63-4416-b70b-305b17fe7385/volumes" Nov 27 11:42:31 crc kubenswrapper[4796]: I1127 11:42:31.880451 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:42:31 crc kubenswrapper[4796]: I1127 11:42:31.881070 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:43:01 crc kubenswrapper[4796]: I1127 11:43:01.879197 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:43:01 crc kubenswrapper[4796]: I1127 11:43:01.880001 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:43:01 crc kubenswrapper[4796]: I1127 11:43:01.880075 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:43:01 crc kubenswrapper[4796]: I1127 11:43:01.881000 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2bf926c9e1c204e1f7eb84bb17f79b027089b97e9942f906c34dda1a62f3cde"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:43:01 crc kubenswrapper[4796]: I1127 11:43:01.881111 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://b2bf926c9e1c204e1f7eb84bb17f79b027089b97e9942f906c34dda1a62f3cde" gracePeriod=600 Nov 27 11:43:02 crc kubenswrapper[4796]: I1127 11:43:02.390977 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="b2bf926c9e1c204e1f7eb84bb17f79b027089b97e9942f906c34dda1a62f3cde" exitCode=0 Nov 27 11:43:02 crc kubenswrapper[4796]: I1127 11:43:02.391046 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"b2bf926c9e1c204e1f7eb84bb17f79b027089b97e9942f906c34dda1a62f3cde"} Nov 27 11:43:02 crc kubenswrapper[4796]: I1127 11:43:02.391337 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"59d4cd23cae1003b3105c73211942a6903e334d6855a0f24850f7efa7e3d1574"} Nov 27 11:43:02 crc kubenswrapper[4796]: I1127 11:43:02.391357 4796 scope.go:117] "RemoveContainer" containerID="cfd84707341303f4e3c8588f23d09e0d87a32dbd7fe43f86a649e6799d5a18a9" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.157145 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z"] Nov 27 11:45:00 crc kubenswrapper[4796]: E1127 11:45:00.158174 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="extract-content" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158199 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="extract-content" Nov 27 11:45:00 crc kubenswrapper[4796]: E1127 11:45:00.158218 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="extract-utilities" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158232 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="extract-utilities" Nov 27 11:45:00 crc kubenswrapper[4796]: E1127 11:45:00.158262 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="extract-content" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158308 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="extract-content" Nov 27 11:45:00 crc kubenswrapper[4796]: E1127 11:45:00.158327 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="registry-server" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158340 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="registry-server" Nov 27 11:45:00 crc kubenswrapper[4796]: E1127 11:45:00.158355 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="extract-utilities" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158368 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="extract-utilities" Nov 27 11:45:00 crc kubenswrapper[4796]: E1127 11:45:00.158392 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="registry-server" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158405 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="registry-server" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158601 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb789a35-ba63-4416-b70b-305b17fe7385" containerName="registry-server" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.158630 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a794975-9d80-47c9-bf8b-b26b71afa054" containerName="registry-server" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.159255 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.164887 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.165037 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.165954 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z"] Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.326670 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b06300-6722-4ac8-b033-4781f968e40f-secret-volume\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.326816 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b06300-6722-4ac8-b033-4781f968e40f-config-volume\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.326965 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vjbl\" (UniqueName: \"kubernetes.io/projected/b2b06300-6722-4ac8-b033-4781f968e40f-kube-api-access-7vjbl\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.428971 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b06300-6722-4ac8-b033-4781f968e40f-secret-volume\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.429385 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b06300-6722-4ac8-b033-4781f968e40f-config-volume\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.429603 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vjbl\" (UniqueName: \"kubernetes.io/projected/b2b06300-6722-4ac8-b033-4781f968e40f-kube-api-access-7vjbl\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.431113 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b06300-6722-4ac8-b033-4781f968e40f-config-volume\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.439971 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b06300-6722-4ac8-b033-4781f968e40f-secret-volume\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.449358 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vjbl\" (UniqueName: \"kubernetes.io/projected/b2b06300-6722-4ac8-b033-4781f968e40f-kube-api-access-7vjbl\") pod \"collect-profiles-29404065-kfm5z\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.521194 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:00 crc kubenswrapper[4796]: I1127 11:45:00.778328 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z"] Nov 27 11:45:01 crc kubenswrapper[4796]: I1127 11:45:01.212677 4796 generic.go:334] "Generic (PLEG): container finished" podID="b2b06300-6722-4ac8-b033-4781f968e40f" containerID="d25589668379b9394e15cb36d64f57f7e46f13696ba3e7700b0682e04e827ca0" exitCode=0 Nov 27 11:45:01 crc kubenswrapper[4796]: I1127 11:45:01.212738 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" event={"ID":"b2b06300-6722-4ac8-b033-4781f968e40f","Type":"ContainerDied","Data":"d25589668379b9394e15cb36d64f57f7e46f13696ba3e7700b0682e04e827ca0"} Nov 27 11:45:01 crc kubenswrapper[4796]: I1127 11:45:01.212990 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" event={"ID":"b2b06300-6722-4ac8-b033-4781f968e40f","Type":"ContainerStarted","Data":"b4235775668defe0c6755ca170c32106c6c1c24f27bd7b6be9f20757eea6db51"} Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.538145 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.666708 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b06300-6722-4ac8-b033-4781f968e40f-secret-volume\") pod \"b2b06300-6722-4ac8-b033-4781f968e40f\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.666820 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b06300-6722-4ac8-b033-4781f968e40f-config-volume\") pod \"b2b06300-6722-4ac8-b033-4781f968e40f\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.666860 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vjbl\" (UniqueName: \"kubernetes.io/projected/b2b06300-6722-4ac8-b033-4781f968e40f-kube-api-access-7vjbl\") pod \"b2b06300-6722-4ac8-b033-4781f968e40f\" (UID: \"b2b06300-6722-4ac8-b033-4781f968e40f\") " Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.667616 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2b06300-6722-4ac8-b033-4781f968e40f-config-volume" (OuterVolumeSpecName: "config-volume") pod "b2b06300-6722-4ac8-b033-4781f968e40f" (UID: "b2b06300-6722-4ac8-b033-4781f968e40f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.674840 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b06300-6722-4ac8-b033-4781f968e40f-kube-api-access-7vjbl" (OuterVolumeSpecName: "kube-api-access-7vjbl") pod "b2b06300-6722-4ac8-b033-4781f968e40f" (UID: "b2b06300-6722-4ac8-b033-4781f968e40f"). InnerVolumeSpecName "kube-api-access-7vjbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.675079 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2b06300-6722-4ac8-b033-4781f968e40f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b2b06300-6722-4ac8-b033-4781f968e40f" (UID: "b2b06300-6722-4ac8-b033-4781f968e40f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.768655 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2b06300-6722-4ac8-b033-4781f968e40f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.768705 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2b06300-6722-4ac8-b033-4781f968e40f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:45:02 crc kubenswrapper[4796]: I1127 11:45:02.768727 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vjbl\" (UniqueName: \"kubernetes.io/projected/b2b06300-6722-4ac8-b033-4781f968e40f-kube-api-access-7vjbl\") on node \"crc\" DevicePath \"\"" Nov 27 11:45:03 crc kubenswrapper[4796]: I1127 11:45:03.234123 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" event={"ID":"b2b06300-6722-4ac8-b033-4781f968e40f","Type":"ContainerDied","Data":"b4235775668defe0c6755ca170c32106c6c1c24f27bd7b6be9f20757eea6db51"} Nov 27 11:45:03 crc kubenswrapper[4796]: I1127 11:45:03.234165 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4235775668defe0c6755ca170c32106c6c1c24f27bd7b6be9f20757eea6db51" Nov 27 11:45:03 crc kubenswrapper[4796]: I1127 11:45:03.234172 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z" Nov 27 11:45:31 crc kubenswrapper[4796]: I1127 11:45:31.880318 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:45:31 crc kubenswrapper[4796]: I1127 11:45:31.880960 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:46:01 crc kubenswrapper[4796]: I1127 11:46:01.879452 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:46:01 crc kubenswrapper[4796]: I1127 11:46:01.880297 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:46:31 crc kubenswrapper[4796]: I1127 11:46:31.880101 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:46:31 crc kubenswrapper[4796]: I1127 11:46:31.880835 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:46:31 crc kubenswrapper[4796]: I1127 11:46:31.880904 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:46:31 crc kubenswrapper[4796]: I1127 11:46:31.881702 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59d4cd23cae1003b3105c73211942a6903e334d6855a0f24850f7efa7e3d1574"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:46:31 crc kubenswrapper[4796]: I1127 11:46:31.881805 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://59d4cd23cae1003b3105c73211942a6903e334d6855a0f24850f7efa7e3d1574" gracePeriod=600 Nov 27 11:46:32 crc kubenswrapper[4796]: I1127 11:46:32.792769 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"59d4cd23cae1003b3105c73211942a6903e334d6855a0f24850f7efa7e3d1574"} Nov 27 11:46:32 crc kubenswrapper[4796]: I1127 11:46:32.792772 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="59d4cd23cae1003b3105c73211942a6903e334d6855a0f24850f7efa7e3d1574" exitCode=0 Nov 27 11:46:32 crc kubenswrapper[4796]: I1127 11:46:32.794099 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"8b7cb903692ced3867395064a71bd00da73fa8daa661bb81331f102ad27ef1c9"} Nov 27 11:46:32 crc kubenswrapper[4796]: I1127 11:46:32.794112 4796 scope.go:117] "RemoveContainer" containerID="b2bf926c9e1c204e1f7eb84bb17f79b027089b97e9942f906c34dda1a62f3cde" Nov 27 11:49:01 crc kubenswrapper[4796]: I1127 11:49:01.879105 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:49:01 crc kubenswrapper[4796]: I1127 11:49:01.879676 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:49:31 crc kubenswrapper[4796]: I1127 11:49:31.878825 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:49:31 crc kubenswrapper[4796]: I1127 11:49:31.879398 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:50:01 crc kubenswrapper[4796]: I1127 11:50:01.879130 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:50:01 crc kubenswrapper[4796]: I1127 11:50:01.879764 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:50:01 crc kubenswrapper[4796]: I1127 11:50:01.879825 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:50:01 crc kubenswrapper[4796]: I1127 11:50:01.880672 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b7cb903692ced3867395064a71bd00da73fa8daa661bb81331f102ad27ef1c9"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:50:01 crc kubenswrapper[4796]: I1127 11:50:01.880765 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://8b7cb903692ced3867395064a71bd00da73fa8daa661bb81331f102ad27ef1c9" gracePeriod=600 Nov 27 11:50:02 crc kubenswrapper[4796]: I1127 11:50:02.175680 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="8b7cb903692ced3867395064a71bd00da73fa8daa661bb81331f102ad27ef1c9" exitCode=0 Nov 27 11:50:02 crc kubenswrapper[4796]: I1127 11:50:02.175766 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"8b7cb903692ced3867395064a71bd00da73fa8daa661bb81331f102ad27ef1c9"} Nov 27 11:50:02 crc kubenswrapper[4796]: I1127 11:50:02.176058 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8"} Nov 27 11:50:02 crc kubenswrapper[4796]: I1127 11:50:02.176087 4796 scope.go:117] "RemoveContainer" containerID="59d4cd23cae1003b3105c73211942a6903e334d6855a0f24850f7efa7e3d1574" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.099577 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd"] Nov 27 11:50:22 crc kubenswrapper[4796]: E1127 11:50:22.100222 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b06300-6722-4ac8-b033-4781f968e40f" containerName="collect-profiles" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.100233 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b06300-6722-4ac8-b033-4781f968e40f" containerName="collect-profiles" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.100344 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b06300-6722-4ac8-b033-4781f968e40f" containerName="collect-profiles" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.101008 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.103845 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.118635 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd"] Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.258387 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slggp\" (UniqueName: \"kubernetes.io/projected/89a6f0f6-192f-43d8-8f81-5c0570369b9d-kube-api-access-slggp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.258477 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.258524 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.360326 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slggp\" (UniqueName: \"kubernetes.io/projected/89a6f0f6-192f-43d8-8f81-5c0570369b9d-kube-api-access-slggp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.360406 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.360434 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.361075 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.361223 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.380899 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slggp\" (UniqueName: \"kubernetes.io/projected/89a6f0f6-192f-43d8-8f81-5c0570369b9d-kube-api-access-slggp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.418701 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:22 crc kubenswrapper[4796]: I1127 11:50:22.600995 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd"] Nov 27 11:50:22 crc kubenswrapper[4796]: W1127 11:50:22.610339 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89a6f0f6_192f_43d8_8f81_5c0570369b9d.slice/crio-a82cc916d9418ea3263c68e0b886cdd6cc72fc7788ebe44a349858935fd4190b WatchSource:0}: Error finding container a82cc916d9418ea3263c68e0b886cdd6cc72fc7788ebe44a349858935fd4190b: Status 404 returned error can't find the container with id a82cc916d9418ea3263c68e0b886cdd6cc72fc7788ebe44a349858935fd4190b Nov 27 11:50:23 crc kubenswrapper[4796]: I1127 11:50:23.317868 4796 generic.go:334] "Generic (PLEG): container finished" podID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerID="522114a7e87ccc30824f54feaca5ef4d3e31e1b78b42c5ac6b5e90cad75ad837" exitCode=0 Nov 27 11:50:23 crc kubenswrapper[4796]: I1127 11:50:23.318171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" event={"ID":"89a6f0f6-192f-43d8-8f81-5c0570369b9d","Type":"ContainerDied","Data":"522114a7e87ccc30824f54feaca5ef4d3e31e1b78b42c5ac6b5e90cad75ad837"} Nov 27 11:50:23 crc kubenswrapper[4796]: I1127 11:50:23.318658 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" event={"ID":"89a6f0f6-192f-43d8-8f81-5c0570369b9d","Type":"ContainerStarted","Data":"a82cc916d9418ea3263c68e0b886cdd6cc72fc7788ebe44a349858935fd4190b"} Nov 27 11:50:23 crc kubenswrapper[4796]: I1127 11:50:23.321896 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:50:31 crc kubenswrapper[4796]: I1127 11:50:31.366135 4796 generic.go:334] "Generic (PLEG): container finished" podID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerID="72d1b0a794001eb48b40ed889cdbdf168f14665eaf701c8d9ea2e6aacf733c89" exitCode=0 Nov 27 11:50:31 crc kubenswrapper[4796]: I1127 11:50:31.366249 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" event={"ID":"89a6f0f6-192f-43d8-8f81-5c0570369b9d","Type":"ContainerDied","Data":"72d1b0a794001eb48b40ed889cdbdf168f14665eaf701c8d9ea2e6aacf733c89"} Nov 27 11:50:32 crc kubenswrapper[4796]: I1127 11:50:32.376558 4796 generic.go:334] "Generic (PLEG): container finished" podID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerID="e6cc0268faf22eae7e23a7ce2676dc93be32666ee923547c0e696e92921238bd" exitCode=0 Nov 27 11:50:32 crc kubenswrapper[4796]: I1127 11:50:32.376605 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" event={"ID":"89a6f0f6-192f-43d8-8f81-5c0570369b9d","Type":"ContainerDied","Data":"e6cc0268faf22eae7e23a7ce2676dc93be32666ee923547c0e696e92921238bd"} Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.608883 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.739371 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-util\") pod \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.739523 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slggp\" (UniqueName: \"kubernetes.io/projected/89a6f0f6-192f-43d8-8f81-5c0570369b9d-kube-api-access-slggp\") pod \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.739558 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-bundle\") pod \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\" (UID: \"89a6f0f6-192f-43d8-8f81-5c0570369b9d\") " Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.740516 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-bundle" (OuterVolumeSpecName: "bundle") pod "89a6f0f6-192f-43d8-8f81-5c0570369b9d" (UID: "89a6f0f6-192f-43d8-8f81-5c0570369b9d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.745402 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a6f0f6-192f-43d8-8f81-5c0570369b9d-kube-api-access-slggp" (OuterVolumeSpecName: "kube-api-access-slggp") pod "89a6f0f6-192f-43d8-8f81-5c0570369b9d" (UID: "89a6f0f6-192f-43d8-8f81-5c0570369b9d"). InnerVolumeSpecName "kube-api-access-slggp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.753672 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-util" (OuterVolumeSpecName: "util") pod "89a6f0f6-192f-43d8-8f81-5c0570369b9d" (UID: "89a6f0f6-192f-43d8-8f81-5c0570369b9d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.849908 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-util\") on node \"crc\" DevicePath \"\"" Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.849966 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slggp\" (UniqueName: \"kubernetes.io/projected/89a6f0f6-192f-43d8-8f81-5c0570369b9d-kube-api-access-slggp\") on node \"crc\" DevicePath \"\"" Nov 27 11:50:33 crc kubenswrapper[4796]: I1127 11:50:33.849984 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89a6f0f6-192f-43d8-8f81-5c0570369b9d-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:50:34 crc kubenswrapper[4796]: I1127 11:50:34.392527 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" event={"ID":"89a6f0f6-192f-43d8-8f81-5c0570369b9d","Type":"ContainerDied","Data":"a82cc916d9418ea3263c68e0b886cdd6cc72fc7788ebe44a349858935fd4190b"} Nov 27 11:50:34 crc kubenswrapper[4796]: I1127 11:50:34.392580 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a82cc916d9418ea3263c68e0b886cdd6cc72fc7788ebe44a349858935fd4190b" Nov 27 11:50:34 crc kubenswrapper[4796]: I1127 11:50:34.392586 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.668967 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw"] Nov 27 11:50:38 crc kubenswrapper[4796]: E1127 11:50:38.669536 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="util" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.669547 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="util" Nov 27 11:50:38 crc kubenswrapper[4796]: E1127 11:50:38.669566 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="extract" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.669572 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="extract" Nov 27 11:50:38 crc kubenswrapper[4796]: E1127 11:50:38.669582 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="pull" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.669588 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="pull" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.669669 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a6f0f6-192f-43d8-8f81-5c0570369b9d" containerName="extract" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.670050 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.671769 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-pns7d" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.672435 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.672783 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.683162 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw"] Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.821880 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqmzt\" (UniqueName: \"kubernetes.io/projected/2e5faff6-d103-4aa7-ae65-77abe90548c5-kube-api-access-hqmzt\") pod \"nmstate-operator-5b5b58f5c8-l6tgw\" (UID: \"2e5faff6-d103-4aa7-ae65-77abe90548c5\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.923331 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqmzt\" (UniqueName: \"kubernetes.io/projected/2e5faff6-d103-4aa7-ae65-77abe90548c5-kube-api-access-hqmzt\") pod \"nmstate-operator-5b5b58f5c8-l6tgw\" (UID: \"2e5faff6-d103-4aa7-ae65-77abe90548c5\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.945729 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqmzt\" (UniqueName: \"kubernetes.io/projected/2e5faff6-d103-4aa7-ae65-77abe90548c5-kube-api-access-hqmzt\") pod \"nmstate-operator-5b5b58f5c8-l6tgw\" (UID: \"2e5faff6-d103-4aa7-ae65-77abe90548c5\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" Nov 27 11:50:38 crc kubenswrapper[4796]: I1127 11:50:38.990358 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" Nov 27 11:50:39 crc kubenswrapper[4796]: I1127 11:50:39.198985 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw"] Nov 27 11:50:39 crc kubenswrapper[4796]: I1127 11:50:39.433897 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" event={"ID":"2e5faff6-d103-4aa7-ae65-77abe90548c5","Type":"ContainerStarted","Data":"fc9dfce0f1cef64ee9bcacbb20eb81ae92b384a9d2885e2f2fba242ab072aeb4"} Nov 27 11:50:41 crc kubenswrapper[4796]: I1127 11:50:41.446721 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" event={"ID":"2e5faff6-d103-4aa7-ae65-77abe90548c5","Type":"ContainerStarted","Data":"39fb691e9e8ea5819b3892fd03576329989307fbd9e04fe99ecdbff514cdfa0f"} Nov 27 11:50:41 crc kubenswrapper[4796]: I1127 11:50:41.468385 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-l6tgw" podStartSLOduration=1.490978525 podStartE2EDuration="3.468360393s" podCreationTimestamp="2025-11-27 11:50:38 +0000 UTC" firstStartedPulling="2025-11-27 11:50:39.212770378 +0000 UTC m=+1556.731089296" lastFinishedPulling="2025-11-27 11:50:41.190152236 +0000 UTC m=+1558.708471164" observedRunningTime="2025-11-27 11:50:41.462138323 +0000 UTC m=+1558.980457241" watchObservedRunningTime="2025-11-27 11:50:41.468360393 +0000 UTC m=+1558.986679321" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.404049 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.405335 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.407181 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-6lkj8" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.420869 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.421763 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.428055 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.446601 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.458108 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vtgx5"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.459128 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.481110 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.541488 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.542358 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.544150 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xpz2g" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.544170 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.546726 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.555698 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573117 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-nmstate-lock\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573158 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/44d16d5c-7e22-4bf4-824e-9d58d750ac7d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-g7rhk\" (UID: \"44d16d5c-7e22-4bf4-824e-9d58d750ac7d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573221 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hc2t\" (UniqueName: \"kubernetes.io/projected/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-kube-api-access-4hc2t\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573242 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfttg\" (UniqueName: \"kubernetes.io/projected/44d16d5c-7e22-4bf4-824e-9d58d750ac7d-kube-api-access-jfttg\") pod \"nmstate-webhook-5f6d4c5ccb-g7rhk\" (UID: \"44d16d5c-7e22-4bf4-824e-9d58d750ac7d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573546 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-dbus-socket\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573608 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd8w9\" (UniqueName: \"kubernetes.io/projected/72f7a9b6-d5a1-4b35-8221-8c7afcaa203c-kube-api-access-xd8w9\") pod \"nmstate-metrics-7f946cbc9-2tbf2\" (UID: \"72f7a9b6-d5a1-4b35-8221-8c7afcaa203c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.573633 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-ovs-socket\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675534 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-nmstate-lock\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675599 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/68dd98a6-a2d9-4a74-81c4-fab3ae549244-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/44d16d5c-7e22-4bf4-824e-9d58d750ac7d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-g7rhk\" (UID: \"44d16d5c-7e22-4bf4-824e-9d58d750ac7d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675691 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hc2t\" (UniqueName: \"kubernetes.io/projected/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-kube-api-access-4hc2t\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675719 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfttg\" (UniqueName: \"kubernetes.io/projected/44d16d5c-7e22-4bf4-824e-9d58d750ac7d-kube-api-access-jfttg\") pod \"nmstate-webhook-5f6d4c5ccb-g7rhk\" (UID: \"44d16d5c-7e22-4bf4-824e-9d58d750ac7d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675740 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-dbus-socket\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675764 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl8l2\" (UniqueName: \"kubernetes.io/projected/68dd98a6-a2d9-4a74-81c4-fab3ae549244-kube-api-access-vl8l2\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675791 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/68dd98a6-a2d9-4a74-81c4-fab3ae549244-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd8w9\" (UniqueName: \"kubernetes.io/projected/72f7a9b6-d5a1-4b35-8221-8c7afcaa203c-kube-api-access-xd8w9\") pod \"nmstate-metrics-7f946cbc9-2tbf2\" (UID: \"72f7a9b6-d5a1-4b35-8221-8c7afcaa203c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675845 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-ovs-socket\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675936 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-ovs-socket\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.675985 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-nmstate-lock\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.676575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-dbus-socket\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.682093 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/44d16d5c-7e22-4bf4-824e-9d58d750ac7d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-g7rhk\" (UID: \"44d16d5c-7e22-4bf4-824e-9d58d750ac7d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.713235 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hc2t\" (UniqueName: \"kubernetes.io/projected/e3cc1a65-8f8a-477c-96f1-3cd385bfa910-kube-api-access-4hc2t\") pod \"nmstate-handler-vtgx5\" (UID: \"e3cc1a65-8f8a-477c-96f1-3cd385bfa910\") " pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.716020 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd8w9\" (UniqueName: \"kubernetes.io/projected/72f7a9b6-d5a1-4b35-8221-8c7afcaa203c-kube-api-access-xd8w9\") pod \"nmstate-metrics-7f946cbc9-2tbf2\" (UID: \"72f7a9b6-d5a1-4b35-8221-8c7afcaa203c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.717313 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfttg\" (UniqueName: \"kubernetes.io/projected/44d16d5c-7e22-4bf4-824e-9d58d750ac7d-kube-api-access-jfttg\") pod \"nmstate-webhook-5f6d4c5ccb-g7rhk\" (UID: \"44d16d5c-7e22-4bf4-824e-9d58d750ac7d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.724096 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.737426 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.776958 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl8l2\" (UniqueName: \"kubernetes.io/projected/68dd98a6-a2d9-4a74-81c4-fab3ae549244-kube-api-access-vl8l2\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.782284 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/68dd98a6-a2d9-4a74-81c4-fab3ae549244-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.782467 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/68dd98a6-a2d9-4a74-81c4-fab3ae549244-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.783704 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/68dd98a6-a2d9-4a74-81c4-fab3ae549244-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.784049 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.786566 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/68dd98a6-a2d9-4a74-81c4-fab3ae549244-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.806056 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl8l2\" (UniqueName: \"kubernetes.io/projected/68dd98a6-a2d9-4a74-81c4-fab3ae549244-kube-api-access-vl8l2\") pod \"nmstate-console-plugin-7fbb5f6569-g8jfz\" (UID: \"68dd98a6-a2d9-4a74-81c4-fab3ae549244\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.813778 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-556d96667b-rf7m4"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.815159 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.823903 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-556d96667b-rf7m4"] Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.858352 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.885762 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-oauth-config\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.886108 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-config\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.886175 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-serving-cert\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.886203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-service-ca\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.886223 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-oauth-serving-cert\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.886242 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xppk\" (UniqueName: \"kubernetes.io/projected/5299efd9-08ea-4be5-9f02-6e6e55ec418f-kube-api-access-7xppk\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.886260 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-trusted-ca-bundle\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: W1127 11:50:42.915715 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3cc1a65_8f8a_477c_96f1_3cd385bfa910.slice/crio-2c7bbac1470526fc6ce4b6a18b03233ab405cc772dd1e6a3110a5d67e2eb530e WatchSource:0}: Error finding container 2c7bbac1470526fc6ce4b6a18b03233ab405cc772dd1e6a3110a5d67e2eb530e: Status 404 returned error can't find the container with id 2c7bbac1470526fc6ce4b6a18b03233ab405cc772dd1e6a3110a5d67e2eb530e Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.989949 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-service-ca\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.990777 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-service-ca\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.990831 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-oauth-serving-cert\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.990856 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xppk\" (UniqueName: \"kubernetes.io/projected/5299efd9-08ea-4be5-9f02-6e6e55ec418f-kube-api-access-7xppk\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.990873 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-trusted-ca-bundle\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.990904 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-oauth-config\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.990930 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-config\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.991008 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-serving-cert\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.992347 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-trusted-ca-bundle\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.992547 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-config\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:42 crc kubenswrapper[4796]: I1127 11:50:42.992641 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5299efd9-08ea-4be5-9f02-6e6e55ec418f-oauth-serving-cert\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:42.998107 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-oauth-config\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:42.998614 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5299efd9-08ea-4be5-9f02-6e6e55ec418f-console-serving-cert\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.007827 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xppk\" (UniqueName: \"kubernetes.io/projected/5299efd9-08ea-4be5-9f02-6e6e55ec418f-kube-api-access-7xppk\") pod \"console-556d96667b-rf7m4\" (UID: \"5299efd9-08ea-4be5-9f02-6e6e55ec418f\") " pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.136508 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz"] Nov 27 11:50:43 crc kubenswrapper[4796]: W1127 11:50:43.139218 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68dd98a6_a2d9_4a74_81c4_fab3ae549244.slice/crio-b8f363a41171ffbcaabcf4c1555fe811dda0e2d88e139ae8a5776326c0972d53 WatchSource:0}: Error finding container b8f363a41171ffbcaabcf4c1555fe811dda0e2d88e139ae8a5776326c0972d53: Status 404 returned error can't find the container with id b8f363a41171ffbcaabcf4c1555fe811dda0e2d88e139ae8a5776326c0972d53 Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.158921 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.228391 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2"] Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.308321 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk"] Nov 27 11:50:43 crc kubenswrapper[4796]: W1127 11:50:43.317866 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44d16d5c_7e22_4bf4_824e_9d58d750ac7d.slice/crio-5962c8af4cbe55845d4e2af61ddec4477d7fa79e557702ebea4def48b5a2f3e8 WatchSource:0}: Error finding container 5962c8af4cbe55845d4e2af61ddec4477d7fa79e557702ebea4def48b5a2f3e8: Status 404 returned error can't find the container with id 5962c8af4cbe55845d4e2af61ddec4477d7fa79e557702ebea4def48b5a2f3e8 Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.339369 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-556d96667b-rf7m4"] Nov 27 11:50:43 crc kubenswrapper[4796]: W1127 11:50:43.342651 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5299efd9_08ea_4be5_9f02_6e6e55ec418f.slice/crio-7677fe78473dfec04708e8eb026205c901eb16d740ccd88ef9ff99217b923e3f WatchSource:0}: Error finding container 7677fe78473dfec04708e8eb026205c901eb16d740ccd88ef9ff99217b923e3f: Status 404 returned error can't find the container with id 7677fe78473dfec04708e8eb026205c901eb16d740ccd88ef9ff99217b923e3f Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.465865 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" event={"ID":"68dd98a6-a2d9-4a74-81c4-fab3ae549244","Type":"ContainerStarted","Data":"b8f363a41171ffbcaabcf4c1555fe811dda0e2d88e139ae8a5776326c0972d53"} Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.466905 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vtgx5" event={"ID":"e3cc1a65-8f8a-477c-96f1-3cd385bfa910","Type":"ContainerStarted","Data":"2c7bbac1470526fc6ce4b6a18b03233ab405cc772dd1e6a3110a5d67e2eb530e"} Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.467970 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-556d96667b-rf7m4" event={"ID":"5299efd9-08ea-4be5-9f02-6e6e55ec418f","Type":"ContainerStarted","Data":"7677fe78473dfec04708e8eb026205c901eb16d740ccd88ef9ff99217b923e3f"} Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.470351 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" event={"ID":"72f7a9b6-d5a1-4b35-8221-8c7afcaa203c","Type":"ContainerStarted","Data":"f2a1e0db2600d16266b2f0956a1b3158c11449a249ff8d0271e0f64389dff3ff"} Nov 27 11:50:43 crc kubenswrapper[4796]: I1127 11:50:43.471300 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" event={"ID":"44d16d5c-7e22-4bf4-824e-9d58d750ac7d","Type":"ContainerStarted","Data":"5962c8af4cbe55845d4e2af61ddec4477d7fa79e557702ebea4def48b5a2f3e8"} Nov 27 11:50:44 crc kubenswrapper[4796]: I1127 11:50:44.478558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-556d96667b-rf7m4" event={"ID":"5299efd9-08ea-4be5-9f02-6e6e55ec418f","Type":"ContainerStarted","Data":"b9db5fdc44979febb47f87ea488315386020ce40fa05097268750e50fee17586"} Nov 27 11:50:44 crc kubenswrapper[4796]: I1127 11:50:44.504424 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-556d96667b-rf7m4" podStartSLOduration=2.504369293 podStartE2EDuration="2.504369293s" podCreationTimestamp="2025-11-27 11:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:50:44.498699037 +0000 UTC m=+1562.017017955" watchObservedRunningTime="2025-11-27 11:50:44.504369293 +0000 UTC m=+1562.022688211" Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.503666 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" event={"ID":"68dd98a6-a2d9-4a74-81c4-fab3ae549244","Type":"ContainerStarted","Data":"34acbf28356c3ea4cdec8c9bcf85a75c0f09d1c6016c285661277ad8ad4e9242"} Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.507542 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vtgx5" event={"ID":"e3cc1a65-8f8a-477c-96f1-3cd385bfa910","Type":"ContainerStarted","Data":"221fa052b11ad3b2b607e5ede1cf25d0e8fb9eee0316df92f3e50ce839b1c069"} Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.507656 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.509000 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" event={"ID":"72f7a9b6-d5a1-4b35-8221-8c7afcaa203c","Type":"ContainerStarted","Data":"47d01e639e5d337cbf3bdc93c53e0046de19f88e0fc6d599561a0ab57eba5c61"} Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.510410 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" event={"ID":"44d16d5c-7e22-4bf4-824e-9d58d750ac7d","Type":"ContainerStarted","Data":"3312dcc4ae3edcfcf059a6fb93165fcd0c049888b9ff362c9ad3f331764bcb23"} Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.510579 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.524542 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8jfz" podStartSLOduration=1.815262954 podStartE2EDuration="4.524519061s" podCreationTimestamp="2025-11-27 11:50:42 +0000 UTC" firstStartedPulling="2025-11-27 11:50:43.141453402 +0000 UTC m=+1560.659772310" lastFinishedPulling="2025-11-27 11:50:45.850709469 +0000 UTC m=+1563.369028417" observedRunningTime="2025-11-27 11:50:46.519782591 +0000 UTC m=+1564.038101519" watchObservedRunningTime="2025-11-27 11:50:46.524519061 +0000 UTC m=+1564.042837979" Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.545281 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" podStartSLOduration=2.005926606 podStartE2EDuration="4.545240557s" podCreationTimestamp="2025-11-27 11:50:42 +0000 UTC" firstStartedPulling="2025-11-27 11:50:43.322767969 +0000 UTC m=+1560.841086907" lastFinishedPulling="2025-11-27 11:50:45.86208194 +0000 UTC m=+1563.380400858" observedRunningTime="2025-11-27 11:50:46.541552856 +0000 UTC m=+1564.059871804" watchObservedRunningTime="2025-11-27 11:50:46.545240557 +0000 UTC m=+1564.063559475" Nov 27 11:50:46 crc kubenswrapper[4796]: I1127 11:50:46.565813 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vtgx5" podStartSLOduration=1.64689934 podStartE2EDuration="4.565792958s" podCreationTimestamp="2025-11-27 11:50:42 +0000 UTC" firstStartedPulling="2025-11-27 11:50:42.927842493 +0000 UTC m=+1560.446161411" lastFinishedPulling="2025-11-27 11:50:45.846736081 +0000 UTC m=+1563.365055029" observedRunningTime="2025-11-27 11:50:46.564552835 +0000 UTC m=+1564.082871753" watchObservedRunningTime="2025-11-27 11:50:46.565792958 +0000 UTC m=+1564.084111876" Nov 27 11:50:49 crc kubenswrapper[4796]: I1127 11:50:49.528848 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" event={"ID":"72f7a9b6-d5a1-4b35-8221-8c7afcaa203c","Type":"ContainerStarted","Data":"9b3bd8dd2a831fc956f4c77d7c4a2e385bc430f6505cff043ae88c1ad0e45bd4"} Nov 27 11:50:49 crc kubenswrapper[4796]: I1127 11:50:49.552195 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-2tbf2" podStartSLOduration=1.890630603 podStartE2EDuration="7.552170871s" podCreationTimestamp="2025-11-27 11:50:42 +0000 UTC" firstStartedPulling="2025-11-27 11:50:43.241537269 +0000 UTC m=+1560.759856187" lastFinishedPulling="2025-11-27 11:50:48.903077537 +0000 UTC m=+1566.421396455" observedRunningTime="2025-11-27 11:50:49.547831783 +0000 UTC m=+1567.066150721" watchObservedRunningTime="2025-11-27 11:50:49.552170871 +0000 UTC m=+1567.070489789" Nov 27 11:50:52 crc kubenswrapper[4796]: I1127 11:50:52.817959 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vtgx5" Nov 27 11:50:53 crc kubenswrapper[4796]: I1127 11:50:53.159689 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:53 crc kubenswrapper[4796]: I1127 11:50:53.159866 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:53 crc kubenswrapper[4796]: I1127 11:50:53.168917 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:53 crc kubenswrapper[4796]: I1127 11:50:53.563971 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-556d96667b-rf7m4" Nov 27 11:50:53 crc kubenswrapper[4796]: I1127 11:50:53.638816 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bvccg"] Nov 27 11:51:02 crc kubenswrapper[4796]: I1127 11:51:02.747316 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-g7rhk" Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.902015 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2"] Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.905438 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.908503 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.928843 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2"] Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.990546 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.990669 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:16 crc kubenswrapper[4796]: I1127 11:51:16.990699 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h94cl\" (UniqueName: \"kubernetes.io/projected/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-kube-api-access-h94cl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.091662 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.091738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h94cl\" (UniqueName: \"kubernetes.io/projected/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-kube-api-access-h94cl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.091821 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.092732 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.092919 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.119539 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h94cl\" (UniqueName: \"kubernetes.io/projected/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-kube-api-access-h94cl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.254228 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:17 crc kubenswrapper[4796]: I1127 11:51:17.704360 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2"] Nov 27 11:51:18 crc kubenswrapper[4796]: I1127 11:51:18.700003 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-bvccg" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerName="console" containerID="cri-o://f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524" gracePeriod=15 Nov 27 11:51:18 crc kubenswrapper[4796]: I1127 11:51:18.731464 4796 generic.go:334] "Generic (PLEG): container finished" podID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerID="8c90a195c9b0ff7ca3ffd53ac97c0f1be9ea38d5a80bf93d363dce485635d600" exitCode=0 Nov 27 11:51:18 crc kubenswrapper[4796]: I1127 11:51:18.731501 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" event={"ID":"00caecbc-f43a-453c-a3b3-a0bfa7c49f60","Type":"ContainerDied","Data":"8c90a195c9b0ff7ca3ffd53ac97c0f1be9ea38d5a80bf93d363dce485635d600"} Nov 27 11:51:18 crc kubenswrapper[4796]: I1127 11:51:18.731525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" event={"ID":"00caecbc-f43a-453c-a3b3-a0bfa7c49f60","Type":"ContainerStarted","Data":"d4ed08c6951bd9b0f5a2da34b094a18535b814ebfe531e8640b8cc53a5bd5710"} Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.112536 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bvccg_222d84fa-7732-4a9d-b250-e66c95092ad6/console/0.log" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.113024 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222385 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-oauth-serving-cert\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222468 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr7bs\" (UniqueName: \"kubernetes.io/projected/222d84fa-7732-4a9d-b250-e66c95092ad6-kube-api-access-lr7bs\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222522 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-service-ca\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222561 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-serving-cert\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222768 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-oauth-config\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222838 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-trusted-ca-bundle\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.222869 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-console-config\") pod \"222d84fa-7732-4a9d-b250-e66c95092ad6\" (UID: \"222d84fa-7732-4a9d-b250-e66c95092ad6\") " Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.223442 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.224209 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-console-config" (OuterVolumeSpecName: "console-config") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.224344 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-service-ca" (OuterVolumeSpecName: "service-ca") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.225510 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.229994 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.230855 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/222d84fa-7732-4a9d-b250-e66c95092ad6-kube-api-access-lr7bs" (OuterVolumeSpecName: "kube-api-access-lr7bs") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "kube-api-access-lr7bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.233071 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "222d84fa-7732-4a9d-b250-e66c95092ad6" (UID: "222d84fa-7732-4a9d-b250-e66c95092ad6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324610 4796 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324652 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr7bs\" (UniqueName: \"kubernetes.io/projected/222d84fa-7732-4a9d-b250-e66c95092ad6-kube-api-access-lr7bs\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324667 4796 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324678 4796 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324690 4796 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/222d84fa-7732-4a9d-b250-e66c95092ad6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324702 4796 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.324713 4796 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/222d84fa-7732-4a9d-b250-e66c95092ad6-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.741513 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bvccg_222d84fa-7732-4a9d-b250-e66c95092ad6/console/0.log" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.742467 4796 generic.go:334] "Generic (PLEG): container finished" podID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerID="f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524" exitCode=2 Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.742531 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bvccg" event={"ID":"222d84fa-7732-4a9d-b250-e66c95092ad6","Type":"ContainerDied","Data":"f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524"} Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.742582 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bvccg" event={"ID":"222d84fa-7732-4a9d-b250-e66c95092ad6","Type":"ContainerDied","Data":"971fa1309f2fb3f0d861dd96b87f9b97c4eaf1305d13a36f5f0499dca602db76"} Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.742608 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bvccg" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.742624 4796 scope.go:117] "RemoveContainer" containerID="f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.772560 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bvccg"] Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.774155 4796 scope.go:117] "RemoveContainer" containerID="f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524" Nov 27 11:51:19 crc kubenswrapper[4796]: E1127 11:51:19.774868 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524\": container with ID starting with f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524 not found: ID does not exist" containerID="f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.775137 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524"} err="failed to get container status \"f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524\": rpc error: code = NotFound desc = could not find container \"f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524\": container with ID starting with f6a6d2ade2d96459ac2af3d21e56c5088810274ce4059eedd65ab0875c9b0524 not found: ID does not exist" Nov 27 11:51:19 crc kubenswrapper[4796]: I1127 11:51:19.780489 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-bvccg"] Nov 27 11:51:21 crc kubenswrapper[4796]: I1127 11:51:21.581411 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" path="/var/lib/kubelet/pods/222d84fa-7732-4a9d-b250-e66c95092ad6/volumes" Nov 27 11:51:21 crc kubenswrapper[4796]: I1127 11:51:21.771557 4796 generic.go:334] "Generic (PLEG): container finished" podID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerID="adefed9a8ecc8a179e1adb39dcda2c8694b57f4ba02e23a1b4b79362a8f75cbc" exitCode=0 Nov 27 11:51:21 crc kubenswrapper[4796]: I1127 11:51:21.771644 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" event={"ID":"00caecbc-f43a-453c-a3b3-a0bfa7c49f60","Type":"ContainerDied","Data":"adefed9a8ecc8a179e1adb39dcda2c8694b57f4ba02e23a1b4b79362a8f75cbc"} Nov 27 11:51:22 crc kubenswrapper[4796]: I1127 11:51:22.781576 4796 generic.go:334] "Generic (PLEG): container finished" podID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerID="0e1cf3e8aec9d612afba8bab9cd7e64622d82c1124e7300e68dd5d079c2720b3" exitCode=0 Nov 27 11:51:22 crc kubenswrapper[4796]: I1127 11:51:22.781664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" event={"ID":"00caecbc-f43a-453c-a3b3-a0bfa7c49f60","Type":"ContainerDied","Data":"0e1cf3e8aec9d612afba8bab9cd7e64622d82c1124e7300e68dd5d079c2720b3"} Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.104901 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.290245 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-bundle\") pod \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.290324 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h94cl\" (UniqueName: \"kubernetes.io/projected/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-kube-api-access-h94cl\") pod \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.290405 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-util\") pod \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\" (UID: \"00caecbc-f43a-453c-a3b3-a0bfa7c49f60\") " Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.291256 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-bundle" (OuterVolumeSpecName: "bundle") pod "00caecbc-f43a-453c-a3b3-a0bfa7c49f60" (UID: "00caecbc-f43a-453c-a3b3-a0bfa7c49f60"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.302038 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-util" (OuterVolumeSpecName: "util") pod "00caecbc-f43a-453c-a3b3-a0bfa7c49f60" (UID: "00caecbc-f43a-453c-a3b3-a0bfa7c49f60"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.303062 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-kube-api-access-h94cl" (OuterVolumeSpecName: "kube-api-access-h94cl") pod "00caecbc-f43a-453c-a3b3-a0bfa7c49f60" (UID: "00caecbc-f43a-453c-a3b3-a0bfa7c49f60"). InnerVolumeSpecName "kube-api-access-h94cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.392023 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.392071 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h94cl\" (UniqueName: \"kubernetes.io/projected/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-kube-api-access-h94cl\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.392088 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/00caecbc-f43a-453c-a3b3-a0bfa7c49f60-util\") on node \"crc\" DevicePath \"\"" Nov 27 11:51:24 crc kubenswrapper[4796]: E1127 11:51:24.639698 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/NetworkManager-dispatcher.service\": RecentStats: unable to find data in memory cache]" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.796664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" event={"ID":"00caecbc-f43a-453c-a3b3-a0bfa7c49f60","Type":"ContainerDied","Data":"d4ed08c6951bd9b0f5a2da34b094a18535b814ebfe531e8640b8cc53a5bd5710"} Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.797045 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4ed08c6951bd9b0f5a2da34b094a18535b814ebfe531e8640b8cc53a5bd5710" Nov 27 11:51:24 crc kubenswrapper[4796]: I1127 11:51:24.796738 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.655622 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf"] Nov 27 11:51:37 crc kubenswrapper[4796]: E1127 11:51:37.656492 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="extract" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.656510 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="extract" Nov 27 11:51:37 crc kubenswrapper[4796]: E1127 11:51:37.656524 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="util" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.656532 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="util" Nov 27 11:51:37 crc kubenswrapper[4796]: E1127 11:51:37.656546 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerName="console" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.656553 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerName="console" Nov 27 11:51:37 crc kubenswrapper[4796]: E1127 11:51:37.656569 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="pull" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.656577 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="pull" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.656710 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="00caecbc-f43a-453c-a3b3-a0bfa7c49f60" containerName="extract" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.656730 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="222d84fa-7732-4a9d-b250-e66c95092ad6" containerName="console" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.657175 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.662623 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.662657 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-trvzp" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.663670 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.663672 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.664926 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.676656 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf"] Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.759740 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-webhook-cert\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.759811 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5j69\" (UniqueName: \"kubernetes.io/projected/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-kube-api-access-l5j69\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.759863 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-apiservice-cert\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.860876 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-apiservice-cert\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.860969 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-webhook-cert\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.861015 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5j69\" (UniqueName: \"kubernetes.io/projected/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-kube-api-access-l5j69\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.868089 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-apiservice-cert\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.869998 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-webhook-cert\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.898769 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5j69\" (UniqueName: \"kubernetes.io/projected/e27fc43a-060d-4f5f-9f52-4f75c9e4b651-kube-api-access-l5j69\") pod \"metallb-operator-controller-manager-68bdcff486-mtshf\" (UID: \"e27fc43a-060d-4f5f-9f52-4f75c9e4b651\") " pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:37 crc kubenswrapper[4796]: I1127 11:51:37.978422 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.119749 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck"] Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.120820 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.125401 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-cdmdg" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.126186 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.126376 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.143872 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck"] Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.268385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4g8n\" (UniqueName: \"kubernetes.io/projected/c1a62d95-1792-4779-8f23-da2e94806d72-kube-api-access-t4g8n\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.268440 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1a62d95-1792-4779-8f23-da2e94806d72-webhook-cert\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.268588 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1a62d95-1792-4779-8f23-da2e94806d72-apiservice-cert\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.370243 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1a62d95-1792-4779-8f23-da2e94806d72-apiservice-cert\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.370615 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4g8n\" (UniqueName: \"kubernetes.io/projected/c1a62d95-1792-4779-8f23-da2e94806d72-kube-api-access-t4g8n\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.370672 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1a62d95-1792-4779-8f23-da2e94806d72-webhook-cert\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.376849 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1a62d95-1792-4779-8f23-da2e94806d72-apiservice-cert\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.386425 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1a62d95-1792-4779-8f23-da2e94806d72-webhook-cert\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.399650 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4g8n\" (UniqueName: \"kubernetes.io/projected/c1a62d95-1792-4779-8f23-da2e94806d72-kube-api-access-t4g8n\") pod \"metallb-operator-webhook-server-57ffb79475-qkvck\" (UID: \"c1a62d95-1792-4779-8f23-da2e94806d72\") " pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.476481 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.546247 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf"] Nov 27 11:51:38 crc kubenswrapper[4796]: W1127 11:51:38.562402 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode27fc43a_060d_4f5f_9f52_4f75c9e4b651.slice/crio-9a56d463ccbcdb79732c2ba10923e2ac1168604a625b77e65e7187c11d7601d5 WatchSource:0}: Error finding container 9a56d463ccbcdb79732c2ba10923e2ac1168604a625b77e65e7187c11d7601d5: Status 404 returned error can't find the container with id 9a56d463ccbcdb79732c2ba10923e2ac1168604a625b77e65e7187c11d7601d5 Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.686300 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck"] Nov 27 11:51:38 crc kubenswrapper[4796]: W1127 11:51:38.693370 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1a62d95_1792_4779_8f23_da2e94806d72.slice/crio-d8608b006da0fb7d8c1cb2d75b4b133b36a9db9d1f0e0c6737d38cab02820ea0 WatchSource:0}: Error finding container d8608b006da0fb7d8c1cb2d75b4b133b36a9db9d1f0e0c6737d38cab02820ea0: Status 404 returned error can't find the container with id d8608b006da0fb7d8c1cb2d75b4b133b36a9db9d1f0e0c6737d38cab02820ea0 Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.886187 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" event={"ID":"e27fc43a-060d-4f5f-9f52-4f75c9e4b651","Type":"ContainerStarted","Data":"9a56d463ccbcdb79732c2ba10923e2ac1168604a625b77e65e7187c11d7601d5"} Nov 27 11:51:38 crc kubenswrapper[4796]: I1127 11:51:38.887240 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" event={"ID":"c1a62d95-1792-4779-8f23-da2e94806d72","Type":"ContainerStarted","Data":"d8608b006da0fb7d8c1cb2d75b4b133b36a9db9d1f0e0c6737d38cab02820ea0"} Nov 27 11:51:43 crc kubenswrapper[4796]: I1127 11:51:43.918846 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" event={"ID":"e27fc43a-060d-4f5f-9f52-4f75c9e4b651","Type":"ContainerStarted","Data":"d4f02a4a623f796c774dcc63db0dff7822c6f4fb9a63404bfe10fddb16b3a0bf"} Nov 27 11:51:43 crc kubenswrapper[4796]: I1127 11:51:43.919283 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:51:43 crc kubenswrapper[4796]: I1127 11:51:43.921084 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" event={"ID":"c1a62d95-1792-4779-8f23-da2e94806d72","Type":"ContainerStarted","Data":"c664f57e6d369eedf20f72809bd02ec0b3cd7442d9c3d8c3d63ea995835322bf"} Nov 27 11:51:43 crc kubenswrapper[4796]: I1127 11:51:43.921249 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:51:43 crc kubenswrapper[4796]: I1127 11:51:43.948111 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" podStartSLOduration=2.470813198 podStartE2EDuration="6.94808842s" podCreationTimestamp="2025-11-27 11:51:37 +0000 UTC" firstStartedPulling="2025-11-27 11:51:38.571218294 +0000 UTC m=+1616.089537212" lastFinishedPulling="2025-11-27 11:51:43.048493516 +0000 UTC m=+1620.566812434" observedRunningTime="2025-11-27 11:51:43.942408354 +0000 UTC m=+1621.460727272" watchObservedRunningTime="2025-11-27 11:51:43.94808842 +0000 UTC m=+1621.466407338" Nov 27 11:51:43 crc kubenswrapper[4796]: I1127 11:51:43.965337 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" podStartSLOduration=1.58285605 podStartE2EDuration="5.9653163s" podCreationTimestamp="2025-11-27 11:51:38 +0000 UTC" firstStartedPulling="2025-11-27 11:51:38.696062937 +0000 UTC m=+1616.214381845" lastFinishedPulling="2025-11-27 11:51:43.078523177 +0000 UTC m=+1620.596842095" observedRunningTime="2025-11-27 11:51:43.964452207 +0000 UTC m=+1621.482771165" watchObservedRunningTime="2025-11-27 11:51:43.9653163 +0000 UTC m=+1621.483635218" Nov 27 11:51:58 crc kubenswrapper[4796]: I1127 11:51:58.482582 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-57ffb79475-qkvck" Nov 27 11:52:17 crc kubenswrapper[4796]: I1127 11:52:17.981985 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-68bdcff486-mtshf" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.738834 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-4hg2d"] Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.741863 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.743591 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-k2nr5" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.743943 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.744014 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.749960 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-85888"] Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.750846 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.752161 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.770029 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-85888"] Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.824733 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9nd74"] Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.825873 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9nd74" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.828574 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.828892 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.835301 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.835353 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-p8dql" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.839176 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-f8gw9"] Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.840452 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.842714 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.849825 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-f8gw9"] Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923844 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg947\" (UniqueName: \"kubernetes.io/projected/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-kube-api-access-qg947\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923895 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923916 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923937 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2w4r\" (UniqueName: \"kubernetes.io/projected/d6863326-623f-4817-ba2b-65f8b1887f5d-kube-api-access-k2w4r\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923966 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-sockets\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923983 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlt57\" (UniqueName: \"kubernetes.io/projected/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-kube-api-access-dlt57\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.923999 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-metrics-certs\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.924022 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d6863326-623f-4817-ba2b-65f8b1887f5d-metallb-excludel2\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.924045 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-reloader\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.924060 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-conf\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.924077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-startup\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.924097 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-metrics\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:18 crc kubenswrapper[4796]: I1127 11:52:18.924111 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-metrics-certs\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025322 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025370 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f8304148-f799-4ea3-880c-7d7e5b744cef-metrics-certs\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025428 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2w4r\" (UniqueName: \"kubernetes.io/projected/d6863326-623f-4817-ba2b-65f8b1887f5d-kube-api-access-k2w4r\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025448 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8304148-f799-4ea3-880c-7d7e5b744cef-cert\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025472 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-sockets\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025516 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlt57\" (UniqueName: \"kubernetes.io/projected/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-kube-api-access-dlt57\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025535 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-metrics-certs\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025552 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sb27\" (UniqueName: \"kubernetes.io/projected/f8304148-f799-4ea3-880c-7d7e5b744cef-kube-api-access-2sb27\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025577 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d6863326-623f-4817-ba2b-65f8b1887f5d-metallb-excludel2\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025600 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-reloader\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025621 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-conf\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-startup\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025658 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-metrics\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025673 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-metrics-certs\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.025698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg947\" (UniqueName: \"kubernetes.io/projected/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-kube-api-access-qg947\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.026067 4796 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.026107 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist podName:d6863326-623f-4817-ba2b-65f8b1887f5d nodeName:}" failed. No retries permitted until 2025-11-27 11:52:19.526093262 +0000 UTC m=+1657.044412180 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist") pod "speaker-9nd74" (UID: "d6863326-623f-4817-ba2b-65f8b1887f5d") : secret "metallb-memberlist" not found Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.026294 4796 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.026324 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-cert podName:10c4179c-1abe-40a0-a5eb-ea7f70a4fde8 nodeName:}" failed. No retries permitted until 2025-11-27 11:52:19.526315508 +0000 UTC m=+1657.044634426 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-cert") pod "frr-k8s-webhook-server-7fcb986d4-85888" (UID: "10c4179c-1abe-40a0-a5eb-ea7f70a4fde8") : secret "frr-k8s-webhook-server-cert" not found Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.026805 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-reloader\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.026868 4796 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.026899 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-metrics-certs podName:d6863326-623f-4817-ba2b-65f8b1887f5d nodeName:}" failed. No retries permitted until 2025-11-27 11:52:19.526887754 +0000 UTC m=+1657.045206752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-metrics-certs") pod "speaker-9nd74" (UID: "d6863326-623f-4817-ba2b-65f8b1887f5d") : secret "speaker-certs-secret" not found Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.026932 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-metrics\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.026976 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-conf\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.027327 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d6863326-623f-4817-ba2b-65f8b1887f5d-metallb-excludel2\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.027398 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-sockets\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.027454 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-frr-startup\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.033087 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-metrics-certs\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.048935 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg947\" (UniqueName: \"kubernetes.io/projected/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-kube-api-access-qg947\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.054893 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlt57\" (UniqueName: \"kubernetes.io/projected/eb8e31ff-f1a5-4a94-84db-ce9642a796b1-kube-api-access-dlt57\") pod \"frr-k8s-4hg2d\" (UID: \"eb8e31ff-f1a5-4a94-84db-ce9642a796b1\") " pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.067301 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2w4r\" (UniqueName: \"kubernetes.io/projected/d6863326-623f-4817-ba2b-65f8b1887f5d-kube-api-access-k2w4r\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.071177 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.127185 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f8304148-f799-4ea3-880c-7d7e5b744cef-metrics-certs\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.127230 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8304148-f799-4ea3-880c-7d7e5b744cef-cert\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.127291 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sb27\" (UniqueName: \"kubernetes.io/projected/f8304148-f799-4ea3-880c-7d7e5b744cef-kube-api-access-2sb27\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.131373 4796 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.132990 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f8304148-f799-4ea3-880c-7d7e5b744cef-metrics-certs\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.144411 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8304148-f799-4ea3-880c-7d7e5b744cef-cert\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.159587 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sb27\" (UniqueName: \"kubernetes.io/projected/f8304148-f799-4ea3-880c-7d7e5b744cef-kube-api-access-2sb27\") pod \"controller-f8648f98b-f8gw9\" (UID: \"f8304148-f799-4ea3-880c-7d7e5b744cef\") " pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.161786 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.416822 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-f8gw9"] Nov 27 11:52:19 crc kubenswrapper[4796]: W1127 11:52:19.425041 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8304148_f799_4ea3_880c_7d7e5b744cef.slice/crio-b733fc5410939965cc0aa43bb534c1c4e13f1a7f52cc494e2862960d17546957 WatchSource:0}: Error finding container b733fc5410939965cc0aa43bb534c1c4e13f1a7f52cc494e2862960d17546957: Status 404 returned error can't find the container with id b733fc5410939965cc0aa43bb534c1c4e13f1a7f52cc494e2862960d17546957 Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.534341 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-metrics-certs\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.534727 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.534756 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.534873 4796 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 11:52:19 crc kubenswrapper[4796]: E1127 11:52:19.534950 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist podName:d6863326-623f-4817-ba2b-65f8b1887f5d nodeName:}" failed. No retries permitted until 2025-11-27 11:52:20.534932792 +0000 UTC m=+1658.053251710 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist") pod "speaker-9nd74" (UID: "d6863326-623f-4817-ba2b-65f8b1887f5d") : secret "metallb-memberlist" not found Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.540608 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-metrics-certs\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.545569 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10c4179c-1abe-40a0-a5eb-ea7f70a4fde8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-85888\" (UID: \"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.686555 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:19 crc kubenswrapper[4796]: I1127 11:52:19.947296 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-85888"] Nov 27 11:52:19 crc kubenswrapper[4796]: W1127 11:52:19.951835 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c4179c_1abe_40a0_a5eb_ea7f70a4fde8.slice/crio-e7dcf12862487850cb17bfa3826669db1bb9e6ec5d902d41b54f9a39da578dd6 WatchSource:0}: Error finding container e7dcf12862487850cb17bfa3826669db1bb9e6ec5d902d41b54f9a39da578dd6: Status 404 returned error can't find the container with id e7dcf12862487850cb17bfa3826669db1bb9e6ec5d902d41b54f9a39da578dd6 Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.154003 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-f8gw9" event={"ID":"f8304148-f799-4ea3-880c-7d7e5b744cef","Type":"ContainerStarted","Data":"fbaa64c380112da8d36d9ef8d29c04239996e1cd14d3fe30622cce8a567cf8b5"} Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.154082 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-f8gw9" event={"ID":"f8304148-f799-4ea3-880c-7d7e5b744cef","Type":"ContainerStarted","Data":"3a80983a0cd6449971e496a42387b81bba9ac102f4985b953b520bf4c3dec091"} Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.154095 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-f8gw9" event={"ID":"f8304148-f799-4ea3-880c-7d7e5b744cef","Type":"ContainerStarted","Data":"b733fc5410939965cc0aa43bb534c1c4e13f1a7f52cc494e2862960d17546957"} Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.154141 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.155178 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" event={"ID":"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8","Type":"ContainerStarted","Data":"e7dcf12862487850cb17bfa3826669db1bb9e6ec5d902d41b54f9a39da578dd6"} Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.156370 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"28d02c924e55b7b8004ebce64f15bd613c5ea8811a2236f315acdf4a6c267ba5"} Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.177683 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-f8gw9" podStartSLOduration=2.177659294 podStartE2EDuration="2.177659294s" podCreationTimestamp="2025-11-27 11:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:52:20.171190187 +0000 UTC m=+1657.689509105" watchObservedRunningTime="2025-11-27 11:52:20.177659294 +0000 UTC m=+1657.695978232" Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.548009 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.555830 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d6863326-623f-4817-ba2b-65f8b1887f5d-memberlist\") pod \"speaker-9nd74\" (UID: \"d6863326-623f-4817-ba2b-65f8b1887f5d\") " pod="metallb-system/speaker-9nd74" Nov 27 11:52:20 crc kubenswrapper[4796]: I1127 11:52:20.649436 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9nd74" Nov 27 11:52:21 crc kubenswrapper[4796]: I1127 11:52:21.171976 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9nd74" event={"ID":"d6863326-623f-4817-ba2b-65f8b1887f5d","Type":"ContainerStarted","Data":"480fc33c1e5978c8ba980664b7ea33107bd09590974dd6ef006f8d52a5401b27"} Nov 27 11:52:21 crc kubenswrapper[4796]: I1127 11:52:21.172338 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9nd74" event={"ID":"d6863326-623f-4817-ba2b-65f8b1887f5d","Type":"ContainerStarted","Data":"3d8f69399eec032f6e055d19c601d4857cda8ee78284969011e9a8a7ae0cab55"} Nov 27 11:52:22 crc kubenswrapper[4796]: I1127 11:52:22.200136 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9nd74" event={"ID":"d6863326-623f-4817-ba2b-65f8b1887f5d","Type":"ContainerStarted","Data":"7bfa3b6c139c31218d9bf40f468c66931e409e43f42b0ba256f8fc6d6f2e3a52"} Nov 27 11:52:22 crc kubenswrapper[4796]: I1127 11:52:22.201913 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9nd74" Nov 27 11:52:22 crc kubenswrapper[4796]: I1127 11:52:22.223766 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9nd74" podStartSLOduration=4.223740421 podStartE2EDuration="4.223740421s" podCreationTimestamp="2025-11-27 11:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:52:22.220210065 +0000 UTC m=+1659.738528983" watchObservedRunningTime="2025-11-27 11:52:22.223740421 +0000 UTC m=+1659.742059349" Nov 27 11:52:28 crc kubenswrapper[4796]: I1127 11:52:28.239688 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" event={"ID":"10c4179c-1abe-40a0-a5eb-ea7f70a4fde8","Type":"ContainerStarted","Data":"8d344a7c4ee1100d51a0b8a5682ae9bf310e86e88b8191e5449d8a719521a8a7"} Nov 27 11:52:28 crc kubenswrapper[4796]: I1127 11:52:28.240343 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:28 crc kubenswrapper[4796]: I1127 11:52:28.241485 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb8e31ff-f1a5-4a94-84db-ce9642a796b1" containerID="d86256986c140a3459bdb1ae525557be45ebec8698cbbbc4b702dbc51a16ef3a" exitCode=0 Nov 27 11:52:28 crc kubenswrapper[4796]: I1127 11:52:28.241504 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerDied","Data":"d86256986c140a3459bdb1ae525557be45ebec8698cbbbc4b702dbc51a16ef3a"} Nov 27 11:52:28 crc kubenswrapper[4796]: I1127 11:52:28.263007 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" podStartSLOduration=2.801264823 podStartE2EDuration="10.262971645s" podCreationTimestamp="2025-11-27 11:52:18 +0000 UTC" firstStartedPulling="2025-11-27 11:52:19.954135033 +0000 UTC m=+1657.472453951" lastFinishedPulling="2025-11-27 11:52:27.415841855 +0000 UTC m=+1664.934160773" observedRunningTime="2025-11-27 11:52:28.256049275 +0000 UTC m=+1665.774368193" watchObservedRunningTime="2025-11-27 11:52:28.262971645 +0000 UTC m=+1665.781290603" Nov 27 11:52:29 crc kubenswrapper[4796]: I1127 11:52:29.168681 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-f8gw9" Nov 27 11:52:29 crc kubenswrapper[4796]: I1127 11:52:29.249751 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb8e31ff-f1a5-4a94-84db-ce9642a796b1" containerID="10281239de340864cda87aff2d7db950e4a7f3de29c303cfb4aeb235de243069" exitCode=0 Nov 27 11:52:29 crc kubenswrapper[4796]: I1127 11:52:29.249825 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerDied","Data":"10281239de340864cda87aff2d7db950e4a7f3de29c303cfb4aeb235de243069"} Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.096484 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2c6t8"] Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.098962 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.109897 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2c6t8"] Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.181689 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-utilities\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.181740 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rc5z\" (UniqueName: \"kubernetes.io/projected/dd2f3c97-2f35-45b6-8538-50ed521d29d9-kube-api-access-9rc5z\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.181795 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-catalog-content\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.257516 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb8e31ff-f1a5-4a94-84db-ce9642a796b1" containerID="309caa026a2d0dd4c04ce33f590229533d8daac506e94d81a6e4fc520231b838" exitCode=0 Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.257559 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerDied","Data":"309caa026a2d0dd4c04ce33f590229533d8daac506e94d81a6e4fc520231b838"} Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.283233 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-utilities\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.283299 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rc5z\" (UniqueName: \"kubernetes.io/projected/dd2f3c97-2f35-45b6-8538-50ed521d29d9-kube-api-access-9rc5z\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.283354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-catalog-content\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.283837 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-utilities\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.283922 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-catalog-content\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.307497 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rc5z\" (UniqueName: \"kubernetes.io/projected/dd2f3c97-2f35-45b6-8538-50ed521d29d9-kube-api-access-9rc5z\") pod \"redhat-marketplace-2c6t8\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.425390 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.660423 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9nd74" Nov 27 11:52:30 crc kubenswrapper[4796]: I1127 11:52:30.987412 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2c6t8"] Nov 27 11:52:31 crc kubenswrapper[4796]: W1127 11:52:30.999411 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd2f3c97_2f35_45b6_8538_50ed521d29d9.slice/crio-f9aafb69234d39a2f0bb79b3178f08c595244464b7b2cecccc71b8940452bf6b WatchSource:0}: Error finding container f9aafb69234d39a2f0bb79b3178f08c595244464b7b2cecccc71b8940452bf6b: Status 404 returned error can't find the container with id f9aafb69234d39a2f0bb79b3178f08c595244464b7b2cecccc71b8940452bf6b Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.264166 4796 generic.go:334] "Generic (PLEG): container finished" podID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerID="7978de07e6e90d4af9355bef92bf0cdcd78eadec4ddafe6bce248928da0c5fdf" exitCode=0 Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.264383 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2c6t8" event={"ID":"dd2f3c97-2f35-45b6-8538-50ed521d29d9","Type":"ContainerDied","Data":"7978de07e6e90d4af9355bef92bf0cdcd78eadec4ddafe6bce248928da0c5fdf"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.266801 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2c6t8" event={"ID":"dd2f3c97-2f35-45b6-8538-50ed521d29d9","Type":"ContainerStarted","Data":"f9aafb69234d39a2f0bb79b3178f08c595244464b7b2cecccc71b8940452bf6b"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.276725 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"876fe25f67818d20f0cfe241f9a41663af7c5eaf2098f9883a8213bdaeb5a9e9"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.276895 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"6c3521ed3d72a15d1c0ac6fe14d7a1dc00fda6d5a78338d9756cc3fb988b5587"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.276981 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"47eae3aa3e393c13dd506078fa97343f9ab58e8e141249b54558cee5fb8fdaef"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.277066 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"37693e03f348e4e8fc27820116a8146d02772d5804021f03609dfc11e9344e34"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.277145 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"0aade843c59282516a8dcf8175f784664fd687b195ede25496ac77537d5cf0e5"} Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.879355 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:52:31 crc kubenswrapper[4796]: I1127 11:52:31.879502 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:52:32 crc kubenswrapper[4796]: I1127 11:52:32.285282 4796 generic.go:334] "Generic (PLEG): container finished" podID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerID="9dc7b46682487f610eddb18c380878cc0ab14d5e273e620f04a9c77abfb3a036" exitCode=0 Nov 27 11:52:32 crc kubenswrapper[4796]: I1127 11:52:32.285487 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2c6t8" event={"ID":"dd2f3c97-2f35-45b6-8538-50ed521d29d9","Type":"ContainerDied","Data":"9dc7b46682487f610eddb18c380878cc0ab14d5e273e620f04a9c77abfb3a036"} Nov 27 11:52:32 crc kubenswrapper[4796]: I1127 11:52:32.289960 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4hg2d" event={"ID":"eb8e31ff-f1a5-4a94-84db-ce9642a796b1","Type":"ContainerStarted","Data":"1c589fbe4ff6a1bd1530b11dee476f194f52a76e07639a4828be054439a97ed9"} Nov 27 11:52:32 crc kubenswrapper[4796]: I1127 11:52:32.290665 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:32 crc kubenswrapper[4796]: I1127 11:52:32.323888 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-4hg2d" podStartSLOduration=6.167608492 podStartE2EDuration="14.323873103s" podCreationTimestamp="2025-11-27 11:52:18 +0000 UTC" firstStartedPulling="2025-11-27 11:52:19.240205975 +0000 UTC m=+1656.758524893" lastFinishedPulling="2025-11-27 11:52:27.396470586 +0000 UTC m=+1664.914789504" observedRunningTime="2025-11-27 11:52:32.322137466 +0000 UTC m=+1669.840456384" watchObservedRunningTime="2025-11-27 11:52:32.323873103 +0000 UTC m=+1669.842192011" Nov 27 11:52:33 crc kubenswrapper[4796]: I1127 11:52:33.299797 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2c6t8" event={"ID":"dd2f3c97-2f35-45b6-8538-50ed521d29d9","Type":"ContainerStarted","Data":"cb9b11d13c2a4454d2b7d0b534e89b93cd9f56d1b4501463938a43325f415ff3"} Nov 27 11:52:34 crc kubenswrapper[4796]: I1127 11:52:34.071950 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:34 crc kubenswrapper[4796]: I1127 11:52:34.108770 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:34 crc kubenswrapper[4796]: I1127 11:52:34.129851 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2c6t8" podStartSLOduration=2.724620279 podStartE2EDuration="4.129833406s" podCreationTimestamp="2025-11-27 11:52:30 +0000 UTC" firstStartedPulling="2025-11-27 11:52:31.266230229 +0000 UTC m=+1668.784549147" lastFinishedPulling="2025-11-27 11:52:32.671443356 +0000 UTC m=+1670.189762274" observedRunningTime="2025-11-27 11:52:33.32781113 +0000 UTC m=+1670.846130048" watchObservedRunningTime="2025-11-27 11:52:34.129833406 +0000 UTC m=+1671.648152324" Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.898868 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rbwz6"] Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.901123 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.905253 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.906478 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-c8mzk" Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.906810 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.907242 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rbwz6"] Nov 27 11:52:36 crc kubenswrapper[4796]: I1127 11:52:36.976392 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfkkf\" (UniqueName: \"kubernetes.io/projected/037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5-kube-api-access-lfkkf\") pod \"openstack-operator-index-rbwz6\" (UID: \"037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5\") " pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:37 crc kubenswrapper[4796]: I1127 11:52:37.078354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfkkf\" (UniqueName: \"kubernetes.io/projected/037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5-kube-api-access-lfkkf\") pod \"openstack-operator-index-rbwz6\" (UID: \"037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5\") " pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:37 crc kubenswrapper[4796]: I1127 11:52:37.095688 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfkkf\" (UniqueName: \"kubernetes.io/projected/037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5-kube-api-access-lfkkf\") pod \"openstack-operator-index-rbwz6\" (UID: \"037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5\") " pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:37 crc kubenswrapper[4796]: I1127 11:52:37.231620 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:37 crc kubenswrapper[4796]: I1127 11:52:37.464316 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rbwz6"] Nov 27 11:52:37 crc kubenswrapper[4796]: W1127 11:52:37.469447 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod037f8aa1_b5aa_4877_a71b_7ef5fa45e5d5.slice/crio-e86063e557ffffd7ccfd6114dd9ac612477fe03ee620d5e9b4ade2fbad8be6ed WatchSource:0}: Error finding container e86063e557ffffd7ccfd6114dd9ac612477fe03ee620d5e9b4ade2fbad8be6ed: Status 404 returned error can't find the container with id e86063e557ffffd7ccfd6114dd9ac612477fe03ee620d5e9b4ade2fbad8be6ed Nov 27 11:52:38 crc kubenswrapper[4796]: I1127 11:52:38.343571 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rbwz6" event={"ID":"037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5","Type":"ContainerStarted","Data":"e86063e557ffffd7ccfd6114dd9ac612477fe03ee620d5e9b4ade2fbad8be6ed"} Nov 27 11:52:39 crc kubenswrapper[4796]: I1127 11:52:39.692747 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-85888" Nov 27 11:52:40 crc kubenswrapper[4796]: I1127 11:52:40.425898 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:40 crc kubenswrapper[4796]: I1127 11:52:40.426028 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:40 crc kubenswrapper[4796]: I1127 11:52:40.465343 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:41 crc kubenswrapper[4796]: I1127 11:52:41.369239 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rbwz6" event={"ID":"037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5","Type":"ContainerStarted","Data":"9291be14b8ef43ff9e168c1a066a4e77b31d1c2f2058e21eba880bba79f011fd"} Nov 27 11:52:41 crc kubenswrapper[4796]: I1127 11:52:41.392571 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rbwz6" podStartSLOduration=1.903790431 podStartE2EDuration="5.392550398s" podCreationTimestamp="2025-11-27 11:52:36 +0000 UTC" firstStartedPulling="2025-11-27 11:52:37.47106214 +0000 UTC m=+1674.989381068" lastFinishedPulling="2025-11-27 11:52:40.959822107 +0000 UTC m=+1678.478141035" observedRunningTime="2025-11-27 11:52:41.388533968 +0000 UTC m=+1678.906852886" watchObservedRunningTime="2025-11-27 11:52:41.392550398 +0000 UTC m=+1678.910869316" Nov 27 11:52:41 crc kubenswrapper[4796]: I1127 11:52:41.419063 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.498349 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s674l"] Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.499888 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.516783 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s674l"] Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.665589 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-utilities\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.665678 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx2qh\" (UniqueName: \"kubernetes.io/projected/db6d9397-e4ed-4a09-8c5f-a6458b0af360-kube-api-access-sx2qh\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.665748 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-catalog-content\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.767325 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-catalog-content\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.767429 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-utilities\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.767482 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx2qh\" (UniqueName: \"kubernetes.io/projected/db6d9397-e4ed-4a09-8c5f-a6458b0af360-kube-api-access-sx2qh\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.767853 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-catalog-content\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.767951 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-utilities\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.807521 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx2qh\" (UniqueName: \"kubernetes.io/projected/db6d9397-e4ed-4a09-8c5f-a6458b0af360-kube-api-access-sx2qh\") pod \"redhat-operators-s674l\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:42 crc kubenswrapper[4796]: I1127 11:52:42.863650 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:43 crc kubenswrapper[4796]: I1127 11:52:43.268113 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s674l"] Nov 27 11:52:43 crc kubenswrapper[4796]: I1127 11:52:43.383348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerStarted","Data":"5e918ec50fd5fc65ce39b5c2a55d2996f8d810ca13b33b70445dae9449f443b0"} Nov 27 11:52:44 crc kubenswrapper[4796]: I1127 11:52:44.392222 4796 generic.go:334] "Generic (PLEG): container finished" podID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerID="f2ff851a1717c842917902dd657c3226370f2e1f396f69756aac07fb403053c6" exitCode=0 Nov 27 11:52:44 crc kubenswrapper[4796]: I1127 11:52:44.392313 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerDied","Data":"f2ff851a1717c842917902dd657c3226370f2e1f396f69756aac07fb403053c6"} Nov 27 11:52:45 crc kubenswrapper[4796]: I1127 11:52:45.290086 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2c6t8"] Nov 27 11:52:45 crc kubenswrapper[4796]: I1127 11:52:45.290795 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2c6t8" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="registry-server" containerID="cri-o://cb9b11d13c2a4454d2b7d0b534e89b93cd9f56d1b4501463938a43325f415ff3" gracePeriod=2 Nov 27 11:52:46 crc kubenswrapper[4796]: I1127 11:52:46.409074 4796 generic.go:334] "Generic (PLEG): container finished" podID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerID="cb9b11d13c2a4454d2b7d0b534e89b93cd9f56d1b4501463938a43325f415ff3" exitCode=0 Nov 27 11:52:46 crc kubenswrapper[4796]: I1127 11:52:46.409133 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2c6t8" event={"ID":"dd2f3c97-2f35-45b6-8538-50ed521d29d9","Type":"ContainerDied","Data":"cb9b11d13c2a4454d2b7d0b534e89b93cd9f56d1b4501463938a43325f415ff3"} Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.233387 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.233459 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.283900 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.446904 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-rbwz6" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.678622 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.846583 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rc5z\" (UniqueName: \"kubernetes.io/projected/dd2f3c97-2f35-45b6-8538-50ed521d29d9-kube-api-access-9rc5z\") pod \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.846670 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-utilities\") pod \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.846750 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-catalog-content\") pod \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\" (UID: \"dd2f3c97-2f35-45b6-8538-50ed521d29d9\") " Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.847528 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-utilities" (OuterVolumeSpecName: "utilities") pod "dd2f3c97-2f35-45b6-8538-50ed521d29d9" (UID: "dd2f3c97-2f35-45b6-8538-50ed521d29d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.854997 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2f3c97-2f35-45b6-8538-50ed521d29d9-kube-api-access-9rc5z" (OuterVolumeSpecName: "kube-api-access-9rc5z") pod "dd2f3c97-2f35-45b6-8538-50ed521d29d9" (UID: "dd2f3c97-2f35-45b6-8538-50ed521d29d9"). InnerVolumeSpecName "kube-api-access-9rc5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.863129 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd2f3c97-2f35-45b6-8538-50ed521d29d9" (UID: "dd2f3c97-2f35-45b6-8538-50ed521d29d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.948252 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rc5z\" (UniqueName: \"kubernetes.io/projected/dd2f3c97-2f35-45b6-8538-50ed521d29d9-kube-api-access-9rc5z\") on node \"crc\" DevicePath \"\"" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.948658 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:52:47 crc kubenswrapper[4796]: I1127 11:52:47.948789 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd2f3c97-2f35-45b6-8538-50ed521d29d9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.430026 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2c6t8" event={"ID":"dd2f3c97-2f35-45b6-8538-50ed521d29d9","Type":"ContainerDied","Data":"f9aafb69234d39a2f0bb79b3178f08c595244464b7b2cecccc71b8940452bf6b"} Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.430080 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2c6t8" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.430101 4796 scope.go:117] "RemoveContainer" containerID="cb9b11d13c2a4454d2b7d0b534e89b93cd9f56d1b4501463938a43325f415ff3" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.435371 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerStarted","Data":"2a0f3b84f9bc695d8f8435adc1f1f69e5086b95fb07269a7af1f55fa4627d20b"} Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.453539 4796 scope.go:117] "RemoveContainer" containerID="9dc7b46682487f610eddb18c380878cc0ab14d5e273e620f04a9c77abfb3a036" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.484501 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2c6t8"] Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.488053 4796 scope.go:117] "RemoveContainer" containerID="7978de07e6e90d4af9355bef92bf0cdcd78eadec4ddafe6bce248928da0c5fdf" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.492821 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2c6t8"] Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.732630 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7"] Nov 27 11:52:48 crc kubenswrapper[4796]: E1127 11:52:48.732869 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="extract-utilities" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.732881 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="extract-utilities" Nov 27 11:52:48 crc kubenswrapper[4796]: E1127 11:52:48.732892 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="registry-server" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.732898 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="registry-server" Nov 27 11:52:48 crc kubenswrapper[4796]: E1127 11:52:48.732914 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="extract-content" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.732921 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="extract-content" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.733036 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" containerName="registry-server" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.733832 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.737433 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5w9ml" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.746059 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7"] Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.863936 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-util\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.864042 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-bundle\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.864076 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvwt6\" (UniqueName: \"kubernetes.io/projected/94131852-2ba2-45de-a5d4-da20c127dd6d-kube-api-access-kvwt6\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.965563 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-util\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.965663 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-bundle\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.965701 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvwt6\" (UniqueName: \"kubernetes.io/projected/94131852-2ba2-45de-a5d4-da20c127dd6d-kube-api-access-kvwt6\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.966152 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-util\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.966449 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-bundle\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:48 crc kubenswrapper[4796]: I1127 11:52:48.989490 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvwt6\" (UniqueName: \"kubernetes.io/projected/94131852-2ba2-45de-a5d4-da20c127dd6d-kube-api-access-kvwt6\") pod \"53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.055627 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.075560 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-4hg2d" Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.295769 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7"] Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.446064 4796 generic.go:334] "Generic (PLEG): container finished" podID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerID="2a0f3b84f9bc695d8f8435adc1f1f69e5086b95fb07269a7af1f55fa4627d20b" exitCode=0 Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.446123 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerDied","Data":"2a0f3b84f9bc695d8f8435adc1f1f69e5086b95fb07269a7af1f55fa4627d20b"} Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.448930 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" event={"ID":"94131852-2ba2-45de-a5d4-da20c127dd6d","Type":"ContainerStarted","Data":"2d1fd9e4548640ed172d8f8cb8d23979a965e6b2658ce0d9ec1699b9d6181fed"} Nov 27 11:52:49 crc kubenswrapper[4796]: I1127 11:52:49.579019 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd2f3c97-2f35-45b6-8538-50ed521d29d9" path="/var/lib/kubelet/pods/dd2f3c97-2f35-45b6-8538-50ed521d29d9/volumes" Nov 27 11:52:50 crc kubenswrapper[4796]: I1127 11:52:50.456903 4796 generic.go:334] "Generic (PLEG): container finished" podID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerID="494e04d7b2ea81dbd31e3f1f13c634c6595b345880de82d0808b4bfcb9feda92" exitCode=0 Nov 27 11:52:50 crc kubenswrapper[4796]: I1127 11:52:50.456951 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" event={"ID":"94131852-2ba2-45de-a5d4-da20c127dd6d","Type":"ContainerDied","Data":"494e04d7b2ea81dbd31e3f1f13c634c6595b345880de82d0808b4bfcb9feda92"} Nov 27 11:52:50 crc kubenswrapper[4796]: I1127 11:52:50.460067 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerStarted","Data":"8608b9dcab162c60a36f07a1bcb5ad5bf7240c2d2e84b281a88ca715d416dd26"} Nov 27 11:52:51 crc kubenswrapper[4796]: I1127 11:52:51.468344 4796 generic.go:334] "Generic (PLEG): container finished" podID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerID="d6850543379be1549492cda241c45abcb499fea2841b422d06449ef5cda54309" exitCode=0 Nov 27 11:52:51 crc kubenswrapper[4796]: I1127 11:52:51.468395 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" event={"ID":"94131852-2ba2-45de-a5d4-da20c127dd6d","Type":"ContainerDied","Data":"d6850543379be1549492cda241c45abcb499fea2841b422d06449ef5cda54309"} Nov 27 11:52:51 crc kubenswrapper[4796]: I1127 11:52:51.503838 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s674l" podStartSLOduration=4.046333635 podStartE2EDuration="9.501927892s" podCreationTimestamp="2025-11-27 11:52:42 +0000 UTC" firstStartedPulling="2025-11-27 11:52:44.396497361 +0000 UTC m=+1681.914816289" lastFinishedPulling="2025-11-27 11:52:49.852091618 +0000 UTC m=+1687.370410546" observedRunningTime="2025-11-27 11:52:50.495009205 +0000 UTC m=+1688.013328123" watchObservedRunningTime="2025-11-27 11:52:51.501927892 +0000 UTC m=+1689.020246810" Nov 27 11:52:52 crc kubenswrapper[4796]: I1127 11:52:52.478027 4796 generic.go:334] "Generic (PLEG): container finished" podID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerID="01bf7bd3d9b948d6009e21c0e05eba4f19ac01240ad445adee56ec2c172ee8b2" exitCode=0 Nov 27 11:52:52 crc kubenswrapper[4796]: I1127 11:52:52.478073 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" event={"ID":"94131852-2ba2-45de-a5d4-da20c127dd6d","Type":"ContainerDied","Data":"01bf7bd3d9b948d6009e21c0e05eba4f19ac01240ad445adee56ec2c172ee8b2"} Nov 27 11:52:52 crc kubenswrapper[4796]: I1127 11:52:52.863937 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:52 crc kubenswrapper[4796]: I1127 11:52:52.864630 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.761371 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.919623 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s674l" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="registry-server" probeResult="failure" output=< Nov 27 11:52:53 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 11:52:53 crc kubenswrapper[4796]: > Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.944625 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvwt6\" (UniqueName: \"kubernetes.io/projected/94131852-2ba2-45de-a5d4-da20c127dd6d-kube-api-access-kvwt6\") pod \"94131852-2ba2-45de-a5d4-da20c127dd6d\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.944854 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-bundle\") pod \"94131852-2ba2-45de-a5d4-da20c127dd6d\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.944933 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-util\") pod \"94131852-2ba2-45de-a5d4-da20c127dd6d\" (UID: \"94131852-2ba2-45de-a5d4-da20c127dd6d\") " Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.945722 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-bundle" (OuterVolumeSpecName: "bundle") pod "94131852-2ba2-45de-a5d4-da20c127dd6d" (UID: "94131852-2ba2-45de-a5d4-da20c127dd6d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.951060 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94131852-2ba2-45de-a5d4-da20c127dd6d-kube-api-access-kvwt6" (OuterVolumeSpecName: "kube-api-access-kvwt6") pod "94131852-2ba2-45de-a5d4-da20c127dd6d" (UID: "94131852-2ba2-45de-a5d4-da20c127dd6d"). InnerVolumeSpecName "kube-api-access-kvwt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:52:53 crc kubenswrapper[4796]: I1127 11:52:53.979046 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-util" (OuterVolumeSpecName: "util") pod "94131852-2ba2-45de-a5d4-da20c127dd6d" (UID: "94131852-2ba2-45de-a5d4-da20c127dd6d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:52:54 crc kubenswrapper[4796]: I1127 11:52:54.046948 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-util\") on node \"crc\" DevicePath \"\"" Nov 27 11:52:54 crc kubenswrapper[4796]: I1127 11:52:54.047002 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvwt6\" (UniqueName: \"kubernetes.io/projected/94131852-2ba2-45de-a5d4-da20c127dd6d-kube-api-access-kvwt6\") on node \"crc\" DevicePath \"\"" Nov 27 11:52:54 crc kubenswrapper[4796]: I1127 11:52:54.047027 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94131852-2ba2-45de-a5d4-da20c127dd6d-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:52:54 crc kubenswrapper[4796]: I1127 11:52:54.498567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" event={"ID":"94131852-2ba2-45de-a5d4-da20c127dd6d","Type":"ContainerDied","Data":"2d1fd9e4548640ed172d8f8cb8d23979a965e6b2658ce0d9ec1699b9d6181fed"} Nov 27 11:52:54 crc kubenswrapper[4796]: I1127 11:52:54.498883 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d1fd9e4548640ed172d8f8cb8d23979a965e6b2658ce0d9ec1699b9d6181fed" Nov 27 11:52:54 crc kubenswrapper[4796]: I1127 11:52:54.498664 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.861222 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q"] Nov 27 11:52:57 crc kubenswrapper[4796]: E1127 11:52:57.861761 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="extract" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.861777 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="extract" Nov 27 11:52:57 crc kubenswrapper[4796]: E1127 11:52:57.861794 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="util" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.861802 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="util" Nov 27 11:52:57 crc kubenswrapper[4796]: E1127 11:52:57.861813 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="pull" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.861822 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="pull" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.861932 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="94131852-2ba2-45de-a5d4-da20c127dd6d" containerName="extract" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.862401 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.868142 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-89pll" Nov 27 11:52:57 crc kubenswrapper[4796]: I1127 11:52:57.903109 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q"] Nov 27 11:52:58 crc kubenswrapper[4796]: I1127 11:52:58.042156 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5s4j\" (UniqueName: \"kubernetes.io/projected/b8c1db68-d88e-406c-a551-bfb6fe0df332-kube-api-access-f5s4j\") pod \"openstack-operator-controller-operator-68f75d5585-g6s7q\" (UID: \"b8c1db68-d88e-406c-a551-bfb6fe0df332\") " pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:52:58 crc kubenswrapper[4796]: I1127 11:52:58.143707 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5s4j\" (UniqueName: \"kubernetes.io/projected/b8c1db68-d88e-406c-a551-bfb6fe0df332-kube-api-access-f5s4j\") pod \"openstack-operator-controller-operator-68f75d5585-g6s7q\" (UID: \"b8c1db68-d88e-406c-a551-bfb6fe0df332\") " pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:52:58 crc kubenswrapper[4796]: I1127 11:52:58.183519 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5s4j\" (UniqueName: \"kubernetes.io/projected/b8c1db68-d88e-406c-a551-bfb6fe0df332-kube-api-access-f5s4j\") pod \"openstack-operator-controller-operator-68f75d5585-g6s7q\" (UID: \"b8c1db68-d88e-406c-a551-bfb6fe0df332\") " pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:52:58 crc kubenswrapper[4796]: I1127 11:52:58.186540 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:52:58 crc kubenswrapper[4796]: I1127 11:52:58.693523 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q"] Nov 27 11:52:58 crc kubenswrapper[4796]: W1127 11:52:58.697932 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8c1db68_d88e_406c_a551_bfb6fe0df332.slice/crio-b2d06239e42c750122b0d4afa14ed83d819a15f00294a6d5e8aace5ba7c80c45 WatchSource:0}: Error finding container b2d06239e42c750122b0d4afa14ed83d819a15f00294a6d5e8aace5ba7c80c45: Status 404 returned error can't find the container with id b2d06239e42c750122b0d4afa14ed83d819a15f00294a6d5e8aace5ba7c80c45 Nov 27 11:52:59 crc kubenswrapper[4796]: I1127 11:52:59.541097 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" event={"ID":"b8c1db68-d88e-406c-a551-bfb6fe0df332","Type":"ContainerStarted","Data":"b2d06239e42c750122b0d4afa14ed83d819a15f00294a6d5e8aace5ba7c80c45"} Nov 27 11:53:01 crc kubenswrapper[4796]: I1127 11:53:01.879858 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:53:01 crc kubenswrapper[4796]: I1127 11:53:01.880300 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:53:02 crc kubenswrapper[4796]: I1127 11:53:02.920902 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:53:02 crc kubenswrapper[4796]: I1127 11:53:02.976823 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:53:03 crc kubenswrapper[4796]: I1127 11:53:03.577431 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:53:03 crc kubenswrapper[4796]: I1127 11:53:03.577626 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" event={"ID":"b8c1db68-d88e-406c-a551-bfb6fe0df332","Type":"ContainerStarted","Data":"6fd087ecedf6aa9cbab4592b05d13ecef587d9b40d60af39f615fbab1d177a86"} Nov 27 11:53:03 crc kubenswrapper[4796]: I1127 11:53:03.672206 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" podStartSLOduration=2.90381381 podStartE2EDuration="6.67215441s" podCreationTimestamp="2025-11-27 11:52:57 +0000 UTC" firstStartedPulling="2025-11-27 11:52:58.70079553 +0000 UTC m=+1696.219114458" lastFinishedPulling="2025-11-27 11:53:02.46913613 +0000 UTC m=+1699.987455058" observedRunningTime="2025-11-27 11:53:03.667963315 +0000 UTC m=+1701.186282243" watchObservedRunningTime="2025-11-27 11:53:03.67215441 +0000 UTC m=+1701.190473438" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.284359 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s674l"] Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.284641 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s674l" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="registry-server" containerID="cri-o://8608b9dcab162c60a36f07a1bcb5ad5bf7240c2d2e84b281a88ca715d416dd26" gracePeriod=2 Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.598179 4796 generic.go:334] "Generic (PLEG): container finished" podID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerID="8608b9dcab162c60a36f07a1bcb5ad5bf7240c2d2e84b281a88ca715d416dd26" exitCode=0 Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.598218 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerDied","Data":"8608b9dcab162c60a36f07a1bcb5ad5bf7240c2d2e84b281a88ca715d416dd26"} Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.692006 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.770417 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-utilities\") pod \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.770523 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-catalog-content\") pod \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.770639 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx2qh\" (UniqueName: \"kubernetes.io/projected/db6d9397-e4ed-4a09-8c5f-a6458b0af360-kube-api-access-sx2qh\") pod \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\" (UID: \"db6d9397-e4ed-4a09-8c5f-a6458b0af360\") " Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.772169 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-utilities" (OuterVolumeSpecName: "utilities") pod "db6d9397-e4ed-4a09-8c5f-a6458b0af360" (UID: "db6d9397-e4ed-4a09-8c5f-a6458b0af360"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.776459 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6d9397-e4ed-4a09-8c5f-a6458b0af360-kube-api-access-sx2qh" (OuterVolumeSpecName: "kube-api-access-sx2qh") pod "db6d9397-e4ed-4a09-8c5f-a6458b0af360" (UID: "db6d9397-e4ed-4a09-8c5f-a6458b0af360"). InnerVolumeSpecName "kube-api-access-sx2qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.873385 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx2qh\" (UniqueName: \"kubernetes.io/projected/db6d9397-e4ed-4a09-8c5f-a6458b0af360-kube-api-access-sx2qh\") on node \"crc\" DevicePath \"\"" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.873439 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.890673 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db6d9397-e4ed-4a09-8c5f-a6458b0af360" (UID: "db6d9397-e4ed-4a09-8c5f-a6458b0af360"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:53:05 crc kubenswrapper[4796]: I1127 11:53:05.975231 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db6d9397-e4ed-4a09-8c5f-a6458b0af360-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.606400 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s674l" event={"ID":"db6d9397-e4ed-4a09-8c5f-a6458b0af360","Type":"ContainerDied","Data":"5e918ec50fd5fc65ce39b5c2a55d2996f8d810ca13b33b70445dae9449f443b0"} Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.606437 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s674l" Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.606822 4796 scope.go:117] "RemoveContainer" containerID="8608b9dcab162c60a36f07a1bcb5ad5bf7240c2d2e84b281a88ca715d416dd26" Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.627883 4796 scope.go:117] "RemoveContainer" containerID="2a0f3b84f9bc695d8f8435adc1f1f69e5086b95fb07269a7af1f55fa4627d20b" Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.647413 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s674l"] Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.653291 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s674l"] Nov 27 11:53:06 crc kubenswrapper[4796]: I1127 11:53:06.660992 4796 scope.go:117] "RemoveContainer" containerID="f2ff851a1717c842917902dd657c3226370f2e1f396f69756aac07fb403053c6" Nov 27 11:53:07 crc kubenswrapper[4796]: I1127 11:53:07.577999 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" path="/var/lib/kubelet/pods/db6d9397-e4ed-4a09-8c5f-a6458b0af360/volumes" Nov 27 11:53:08 crc kubenswrapper[4796]: I1127 11:53:08.190673 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-68f75d5585-g6s7q" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.299739 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq"] Nov 27 11:53:26 crc kubenswrapper[4796]: E1127 11:53:26.300674 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="registry-server" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.300691 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="registry-server" Nov 27 11:53:26 crc kubenswrapper[4796]: E1127 11:53:26.300719 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="extract-content" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.300728 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="extract-content" Nov 27 11:53:26 crc kubenswrapper[4796]: E1127 11:53:26.300741 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="extract-utilities" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.300750 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="extract-utilities" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.300879 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6d9397-e4ed-4a09-8c5f-a6458b0af360" containerName="registry-server" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.301640 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.303408 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-cxxlv" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.305422 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.306603 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.308885 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-h2pkt" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.312323 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.325428 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.347662 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-78jwg"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.348752 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.359008 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-5hxrp" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.383475 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.411405 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.436738 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-hcdlf" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.437355 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-78jwg"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.459246 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.467072 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76ldn\" (UniqueName: \"kubernetes.io/projected/efe899c4-0891-480f-9aaa-f8b795706142-kube-api-access-76ldn\") pod \"designate-operator-controller-manager-955677c94-78jwg\" (UID: \"efe899c4-0891-480f-9aaa-f8b795706142\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.467179 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjgf5\" (UniqueName: \"kubernetes.io/projected/ca4033f3-d225-4700-bf24-875462a1a404-kube-api-access-kjgf5\") pod \"barbican-operator-controller-manager-7b64f4fb85-z5fwq\" (UID: \"ca4033f3-d225-4700-bf24-875462a1a404\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.467217 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpvnk\" (UniqueName: \"kubernetes.io/projected/154fc49c-3eca-4d26-836f-ed87d2d78716-kube-api-access-kpvnk\") pod \"cinder-operator-controller-manager-6b7f75547b-lcrbs\" (UID: \"154fc49c-3eca-4d26-836f-ed87d2d78716\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.482937 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.489426 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.497304 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-cxz7x" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.498957 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.500211 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.501632 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-cd5n9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.526554 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.552135 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.567316 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.568233 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjgf5\" (UniqueName: \"kubernetes.io/projected/ca4033f3-d225-4700-bf24-875462a1a404-kube-api-access-kjgf5\") pod \"barbican-operator-controller-manager-7b64f4fb85-z5fwq\" (UID: \"ca4033f3-d225-4700-bf24-875462a1a404\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.568376 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpvnk\" (UniqueName: \"kubernetes.io/projected/154fc49c-3eca-4d26-836f-ed87d2d78716-kube-api-access-kpvnk\") pod \"cinder-operator-controller-manager-6b7f75547b-lcrbs\" (UID: \"154fc49c-3eca-4d26-836f-ed87d2d78716\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.568484 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76ldn\" (UniqueName: \"kubernetes.io/projected/efe899c4-0891-480f-9aaa-f8b795706142-kube-api-access-76ldn\") pod \"designate-operator-controller-manager-955677c94-78jwg\" (UID: \"efe899c4-0891-480f-9aaa-f8b795706142\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.568600 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xnth\" (UniqueName: \"kubernetes.io/projected/0c41793c-bff2-4a6a-877c-489df83fa578-kube-api-access-2xnth\") pod \"glance-operator-controller-manager-589cbd6b5b-zlbmz\" (UID: \"0c41793c-bff2-4a6a-877c-489df83fa578\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.569106 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.577776 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.579062 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.581524 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-q49jk" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.587629 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.589821 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ddrl4" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.593548 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.609467 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.610199 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpvnk\" (UniqueName: \"kubernetes.io/projected/154fc49c-3eca-4d26-836f-ed87d2d78716-kube-api-access-kpvnk\") pod \"cinder-operator-controller-manager-6b7f75547b-lcrbs\" (UID: \"154fc49c-3eca-4d26-836f-ed87d2d78716\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.612865 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.613968 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.616707 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-wgp7x" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.617042 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjgf5\" (UniqueName: \"kubernetes.io/projected/ca4033f3-d225-4700-bf24-875462a1a404-kube-api-access-kjgf5\") pod \"barbican-operator-controller-manager-7b64f4fb85-z5fwq\" (UID: \"ca4033f3-d225-4700-bf24-875462a1a404\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.629820 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.633887 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.635213 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.636044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76ldn\" (UniqueName: \"kubernetes.io/projected/efe899c4-0891-480f-9aaa-f8b795706142-kube-api-access-76ldn\") pod \"designate-operator-controller-manager-955677c94-78jwg\" (UID: \"efe899c4-0891-480f-9aaa-f8b795706142\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.637780 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-l9kbv" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.638094 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.638391 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.639672 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.643539 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7r79d" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.651342 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.663721 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.671989 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c94x9\" (UniqueName: \"kubernetes.io/projected/24bc27f0-611d-4147-9819-6ec0eb012d81-kube-api-access-c94x9\") pod \"horizon-operator-controller-manager-5d494799bf-wkg8j\" (UID: \"24bc27f0-611d-4147-9819-6ec0eb012d81\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672041 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672064 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwqfz\" (UniqueName: \"kubernetes.io/projected/0b1822e4-eff9-4ae6-9d91-46bbef3d090b-kube-api-access-fwqfz\") pod \"ironic-operator-controller-manager-67cb4dc6d4-6fpn9\" (UID: \"0b1822e4-eff9-4ae6-9d91-46bbef3d090b\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672085 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sc7x\" (UniqueName: \"kubernetes.io/projected/2b095ec7-98b1-4e3c-89ed-0bec53da6057-kube-api-access-2sc7x\") pod \"heat-operator-controller-manager-5b77f656f-zqg7w\" (UID: \"2b095ec7-98b1-4e3c-89ed-0bec53da6057\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672103 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xnth\" (UniqueName: \"kubernetes.io/projected/0c41793c-bff2-4a6a-877c-489df83fa578-kube-api-access-2xnth\") pod \"glance-operator-controller-manager-589cbd6b5b-zlbmz\" (UID: \"0c41793c-bff2-4a6a-877c-489df83fa578\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672128 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-846qv\" (UniqueName: \"kubernetes.io/projected/940eec7d-4d26-462b-958b-ec782cf28fdf-kube-api-access-846qv\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672194 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.672220 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.673125 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.689773 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bm62c" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.692712 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.693222 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.707412 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xnth\" (UniqueName: \"kubernetes.io/projected/0c41793c-bff2-4a6a-877c-489df83fa578-kube-api-access-2xnth\") pod \"glance-operator-controller-manager-589cbd6b5b-zlbmz\" (UID: \"0c41793c-bff2-4a6a-877c-489df83fa578\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.711318 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.712354 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.717911 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fbvh6" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.727401 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.732500 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.733558 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.740326 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6wvj7" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.743033 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.770516 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.775033 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-56ch9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.779827 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.784488 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.832515 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866087 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5t6s\" (UniqueName: \"kubernetes.io/projected/d8beead6-b3d4-4f4c-baae-4eaf00fe812a-kube-api-access-l5t6s\") pod \"neutron-operator-controller-manager-6fdcddb789-fsmqp\" (UID: \"d8beead6-b3d4-4f4c-baae-4eaf00fe812a\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866166 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdngf\" (UniqueName: \"kubernetes.io/projected/6858cdee-21cc-4595-905a-3cd4ce0da27c-kube-api-access-jdngf\") pod \"manila-operator-controller-manager-5d499bf58b-ll522\" (UID: \"6858cdee-21cc-4595-905a-3cd4ce0da27c\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866193 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c94x9\" (UniqueName: \"kubernetes.io/projected/24bc27f0-611d-4147-9819-6ec0eb012d81-kube-api-access-c94x9\") pod \"horizon-operator-controller-manager-5d494799bf-wkg8j\" (UID: \"24bc27f0-611d-4147-9819-6ec0eb012d81\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866215 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lpn9\" (UniqueName: \"kubernetes.io/projected/257e6c0f-c3fd-458e-839a-d28ca580fd13-kube-api-access-5lpn9\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-8wjn5\" (UID: \"257e6c0f-c3fd-458e-839a-d28ca580fd13\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866240 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2jwn\" (UniqueName: \"kubernetes.io/projected/5865e45d-8c50-45ea-9fb5-f8a14385de1a-kube-api-access-x2jwn\") pod \"keystone-operator-controller-manager-7b4567c7cf-drlr5\" (UID: \"5865e45d-8c50-45ea-9fb5-f8a14385de1a\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866290 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866318 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwqfz\" (UniqueName: \"kubernetes.io/projected/0b1822e4-eff9-4ae6-9d91-46bbef3d090b-kube-api-access-fwqfz\") pod \"ironic-operator-controller-manager-67cb4dc6d4-6fpn9\" (UID: \"0b1822e4-eff9-4ae6-9d91-46bbef3d090b\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866342 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sc7x\" (UniqueName: \"kubernetes.io/projected/2b095ec7-98b1-4e3c-89ed-0bec53da6057-kube-api-access-2sc7x\") pod \"heat-operator-controller-manager-5b77f656f-zqg7w\" (UID: \"2b095ec7-98b1-4e3c-89ed-0bec53da6057\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866372 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-846qv\" (UniqueName: \"kubernetes.io/projected/940eec7d-4d26-462b-958b-ec782cf28fdf-kube-api-access-846qv\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.866396 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p44l\" (UniqueName: \"kubernetes.io/projected/2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1-kube-api-access-5p44l\") pod \"nova-operator-controller-manager-79556f57fc-6ks4r\" (UID: \"2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:53:26 crc kubenswrapper[4796]: E1127 11:53:26.867175 4796 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:26 crc kubenswrapper[4796]: E1127 11:53:26.867226 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert podName:940eec7d-4d26-462b-958b-ec782cf28fdf nodeName:}" failed. No retries permitted until 2025-11-27 11:53:27.367201209 +0000 UTC m=+1724.885520127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert") pod "infra-operator-controller-manager-57548d458d-zzhd9" (UID: "940eec7d-4d26-462b-958b-ec782cf28fdf") : secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.873454 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.916889 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.918214 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.919111 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwqfz\" (UniqueName: \"kubernetes.io/projected/0b1822e4-eff9-4ae6-9d91-46bbef3d090b-kube-api-access-fwqfz\") pod \"ironic-operator-controller-manager-67cb4dc6d4-6fpn9\" (UID: \"0b1822e4-eff9-4ae6-9d91-46bbef3d090b\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.922368 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-cc45p" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.922974 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-846qv\" (UniqueName: \"kubernetes.io/projected/940eec7d-4d26-462b-958b-ec782cf28fdf-kube-api-access-846qv\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.945488 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sc7x\" (UniqueName: \"kubernetes.io/projected/2b095ec7-98b1-4e3c-89ed-0bec53da6057-kube-api-access-2sc7x\") pod \"heat-operator-controller-manager-5b77f656f-zqg7w\" (UID: \"2b095ec7-98b1-4e3c-89ed-0bec53da6057\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.946001 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c94x9\" (UniqueName: \"kubernetes.io/projected/24bc27f0-611d-4147-9819-6ec0eb012d81-kube-api-access-c94x9\") pod \"horizon-operator-controller-manager-5d494799bf-wkg8j\" (UID: \"24bc27f0-611d-4147-9819-6ec0eb012d81\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971683 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lpn9\" (UniqueName: \"kubernetes.io/projected/257e6c0f-c3fd-458e-839a-d28ca580fd13-kube-api-access-5lpn9\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-8wjn5\" (UID: \"257e6c0f-c3fd-458e-839a-d28ca580fd13\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971715 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdngf\" (UniqueName: \"kubernetes.io/projected/6858cdee-21cc-4595-905a-3cd4ce0da27c-kube-api-access-jdngf\") pod \"manila-operator-controller-manager-5d499bf58b-ll522\" (UID: \"6858cdee-21cc-4595-905a-3cd4ce0da27c\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971739 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2jwn\" (UniqueName: \"kubernetes.io/projected/5865e45d-8c50-45ea-9fb5-f8a14385de1a-kube-api-access-x2jwn\") pod \"keystone-operator-controller-manager-7b4567c7cf-drlr5\" (UID: \"5865e45d-8c50-45ea-9fb5-f8a14385de1a\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971786 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb9tw\" (UniqueName: \"kubernetes.io/projected/b0ba2f52-f61b-47b5-8739-3c248d6e012b-kube-api-access-jb9tw\") pod \"ovn-operator-controller-manager-56897c768d-4xs4w\" (UID: \"b0ba2f52-f61b-47b5-8739-3c248d6e012b\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971823 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9xcg\" (UniqueName: \"kubernetes.io/projected/c2e58505-5b12-41ac-aa51-d857b46b0166-kube-api-access-f9xcg\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971860 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p44l\" (UniqueName: \"kubernetes.io/projected/2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1-kube-api-access-5p44l\") pod \"nova-operator-controller-manager-79556f57fc-6ks4r\" (UID: \"2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971892 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5t6s\" (UniqueName: \"kubernetes.io/projected/d8beead6-b3d4-4f4c-baae-4eaf00fe812a-kube-api-access-l5t6s\") pod \"neutron-operator-controller-manager-6fdcddb789-fsmqp\" (UID: \"d8beead6-b3d4-4f4c-baae-4eaf00fe812a\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971913 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.971967 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqmst\" (UniqueName: \"kubernetes.io/projected/25fd5653-6347-40b2-9d4e-531b58a7e6a3-kube-api-access-gqmst\") pod \"octavia-operator-controller-manager-64cdc6ff96-8lq5v\" (UID: \"25fd5653-6347-40b2-9d4e-531b58a7e6a3\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.991858 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w"] Nov 27 11:53:26 crc kubenswrapper[4796]: I1127 11:53:26.996509 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p44l\" (UniqueName: \"kubernetes.io/projected/2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1-kube-api-access-5p44l\") pod \"nova-operator-controller-manager-79556f57fc-6ks4r\" (UID: \"2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:26.999912 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lpn9\" (UniqueName: \"kubernetes.io/projected/257e6c0f-c3fd-458e-839a-d28ca580fd13-kube-api-access-5lpn9\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-8wjn5\" (UID: \"257e6c0f-c3fd-458e-839a-d28ca580fd13\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.011639 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.016563 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.018063 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.022751 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5t6s\" (UniqueName: \"kubernetes.io/projected/d8beead6-b3d4-4f4c-baae-4eaf00fe812a-kube-api-access-l5t6s\") pod \"neutron-operator-controller-manager-6fdcddb789-fsmqp\" (UID: \"d8beead6-b3d4-4f4c-baae-4eaf00fe812a\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.023838 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jp7tj" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.026753 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-grznn"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.028408 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.029751 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2jwn\" (UniqueName: \"kubernetes.io/projected/5865e45d-8c50-45ea-9fb5-f8a14385de1a-kube-api-access-x2jwn\") pod \"keystone-operator-controller-manager-7b4567c7cf-drlr5\" (UID: \"5865e45d-8c50-45ea-9fb5-f8a14385de1a\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.032012 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdngf\" (UniqueName: \"kubernetes.io/projected/6858cdee-21cc-4595-905a-3cd4ce0da27c-kube-api-access-jdngf\") pod \"manila-operator-controller-manager-5d499bf58b-ll522\" (UID: \"6858cdee-21cc-4595-905a-3cd4ce0da27c\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.032764 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-zx97r" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.039367 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.046595 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-grznn"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.072284 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.072335 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7p6\" (UniqueName: \"kubernetes.io/projected/b72bebe8-8cf0-4e6e-908b-a5095a7a9348-kube-api-access-md7p6\") pod \"swift-operator-controller-manager-d77b94747-grznn\" (UID: \"b72bebe8-8cf0-4e6e-908b-a5095a7a9348\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.072411 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6gft\" (UniqueName: \"kubernetes.io/projected/3189b882-80fe-4198-872d-b8b36f0cc766-kube-api-access-f6gft\") pod \"placement-operator-controller-manager-57988cc5b5-cjtbm\" (UID: \"3189b882-80fe-4198-872d-b8b36f0cc766\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.072449 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqmst\" (UniqueName: \"kubernetes.io/projected/25fd5653-6347-40b2-9d4e-531b58a7e6a3-kube-api-access-gqmst\") pod \"octavia-operator-controller-manager-64cdc6ff96-8lq5v\" (UID: \"25fd5653-6347-40b2-9d4e-531b58a7e6a3\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.072456 4796 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.072490 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb9tw\" (UniqueName: \"kubernetes.io/projected/b0ba2f52-f61b-47b5-8739-3c248d6e012b-kube-api-access-jb9tw\") pod \"ovn-operator-controller-manager-56897c768d-4xs4w\" (UID: \"b0ba2f52-f61b-47b5-8739-3c248d6e012b\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.072535 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9xcg\" (UniqueName: \"kubernetes.io/projected/c2e58505-5b12-41ac-aa51-d857b46b0166-kube-api-access-f9xcg\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.072564 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert podName:c2e58505-5b12-41ac-aa51-d857b46b0166 nodeName:}" failed. No retries permitted until 2025-11-27 11:53:27.572539003 +0000 UTC m=+1725.090857921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" (UID: "c2e58505-5b12-41ac-aa51-d857b46b0166") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.073097 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.074417 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.078520 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5s8dj" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.085948 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.099003 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb9tw\" (UniqueName: \"kubernetes.io/projected/b0ba2f52-f61b-47b5-8739-3c248d6e012b-kube-api-access-jb9tw\") pod \"ovn-operator-controller-manager-56897c768d-4xs4w\" (UID: \"b0ba2f52-f61b-47b5-8739-3c248d6e012b\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.101438 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9xcg\" (UniqueName: \"kubernetes.io/projected/c2e58505-5b12-41ac-aa51-d857b46b0166-kube-api-access-f9xcg\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.103083 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqmst\" (UniqueName: \"kubernetes.io/projected/25fd5653-6347-40b2-9d4e-531b58a7e6a3-kube-api-access-gqmst\") pod \"octavia-operator-controller-manager-64cdc6ff96-8lq5v\" (UID: \"25fd5653-6347-40b2-9d4e-531b58a7e6a3\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.105126 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.106305 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.110554 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4dv6t" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.115447 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.146485 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.150742 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.155437 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.156486 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.159483 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.161871 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.166234 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-m8pm9" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.166914 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.174383 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6gft\" (UniqueName: \"kubernetes.io/projected/3189b882-80fe-4198-872d-b8b36f0cc766-kube-api-access-f6gft\") pod \"placement-operator-controller-manager-57988cc5b5-cjtbm\" (UID: \"3189b882-80fe-4198-872d-b8b36f0cc766\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.174514 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7p6\" (UniqueName: \"kubernetes.io/projected/b72bebe8-8cf0-4e6e-908b-a5095a7a9348-kube-api-access-md7p6\") pod \"swift-operator-controller-manager-d77b94747-grznn\" (UID: \"b72bebe8-8cf0-4e6e-908b-a5095a7a9348\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.203720 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6gft\" (UniqueName: \"kubernetes.io/projected/3189b882-80fe-4198-872d-b8b36f0cc766-kube-api-access-f6gft\") pod \"placement-operator-controller-manager-57988cc5b5-cjtbm\" (UID: \"3189b882-80fe-4198-872d-b8b36f0cc766\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.216818 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.225191 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7p6\" (UniqueName: \"kubernetes.io/projected/b72bebe8-8cf0-4e6e-908b-a5095a7a9348-kube-api-access-md7p6\") pod \"swift-operator-controller-manager-d77b94747-grznn\" (UID: \"b72bebe8-8cf0-4e6e-908b-a5095a7a9348\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.231750 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.232963 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.235494 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-mbdc5" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.235639 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.235863 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.242257 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.273569 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.273805 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.275046 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrvzg\" (UniqueName: \"kubernetes.io/projected/068c68ca-9efb-4a10-b647-9688a007ddf2-kube-api-access-wrvzg\") pod \"telemetry-operator-controller-manager-6f77465788-gr9vn\" (UID: \"068c68ca-9efb-4a10-b647-9688a007ddf2\") " pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.275128 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvqzf\" (UniqueName: \"kubernetes.io/projected/4905ae06-8f67-4beb-93ca-bcd9e1dca63e-kube-api-access-qvqzf\") pod \"watcher-operator-controller-manager-656dcb59d4-88bxr\" (UID: \"4905ae06-8f67-4beb-93ca-bcd9e1dca63e\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.275175 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsbbj\" (UniqueName: \"kubernetes.io/projected/240033a0-6f60-4c73-b2cb-8d4ddb91a66d-kube-api-access-hsbbj\") pod \"test-operator-controller-manager-5cd6c7f4c8-d8k66\" (UID: \"240033a0-6f60-4c73-b2cb-8d4ddb91a66d\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.276427 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.281821 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.282140 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.294324 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.295150 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9xmgf" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.320544 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.348685 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379485 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzjpc\" (UniqueName: \"kubernetes.io/projected/12be3ae6-8893-4998-94bd-fd12c7c6640c-kube-api-access-hzjpc\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379532 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379562 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsbbj\" (UniqueName: \"kubernetes.io/projected/240033a0-6f60-4c73-b2cb-8d4ddb91a66d-kube-api-access-hsbbj\") pod \"test-operator-controller-manager-5cd6c7f4c8-d8k66\" (UID: \"240033a0-6f60-4c73-b2cb-8d4ddb91a66d\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379613 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrvzg\" (UniqueName: \"kubernetes.io/projected/068c68ca-9efb-4a10-b647-9688a007ddf2-kube-api-access-wrvzg\") pod \"telemetry-operator-controller-manager-6f77465788-gr9vn\" (UID: \"068c68ca-9efb-4a10-b647-9688a007ddf2\") " pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379644 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379692 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.379714 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvqzf\" (UniqueName: \"kubernetes.io/projected/4905ae06-8f67-4beb-93ca-bcd9e1dca63e-kube-api-access-qvqzf\") pod \"watcher-operator-controller-manager-656dcb59d4-88bxr\" (UID: \"4905ae06-8f67-4beb-93ca-bcd9e1dca63e\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.380232 4796 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.382127 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert podName:940eec7d-4d26-462b-958b-ec782cf28fdf nodeName:}" failed. No retries permitted until 2025-11-27 11:53:28.382107915 +0000 UTC m=+1725.900426833 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert") pod "infra-operator-controller-manager-57548d458d-zzhd9" (UID: "940eec7d-4d26-462b-958b-ec782cf28fdf") : secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.399038 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrvzg\" (UniqueName: \"kubernetes.io/projected/068c68ca-9efb-4a10-b647-9688a007ddf2-kube-api-access-wrvzg\") pod \"telemetry-operator-controller-manager-6f77465788-gr9vn\" (UID: \"068c68ca-9efb-4a10-b647-9688a007ddf2\") " pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.400849 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsbbj\" (UniqueName: \"kubernetes.io/projected/240033a0-6f60-4c73-b2cb-8d4ddb91a66d-kube-api-access-hsbbj\") pod \"test-operator-controller-manager-5cd6c7f4c8-d8k66\" (UID: \"240033a0-6f60-4c73-b2cb-8d4ddb91a66d\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.404935 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvqzf\" (UniqueName: \"kubernetes.io/projected/4905ae06-8f67-4beb-93ca-bcd9e1dca63e-kube-api-access-qvqzf\") pod \"watcher-operator-controller-manager-656dcb59d4-88bxr\" (UID: \"4905ae06-8f67-4beb-93ca-bcd9e1dca63e\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.418043 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.428887 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.447973 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.480707 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.480778 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzjpc\" (UniqueName: \"kubernetes.io/projected/12be3ae6-8893-4998-94bd-fd12c7c6640c-kube-api-access-hzjpc\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.480813 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.480859 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzz8g\" (UniqueName: \"kubernetes.io/projected/6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb-kube-api-access-tzz8g\") pod \"rabbitmq-cluster-operator-manager-668c99d594-l2v4r\" (UID: \"6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.480909 4796 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.480979 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:27.980960489 +0000 UTC m=+1725.499279407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "metrics-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.481026 4796 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.481074 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:27.981058981 +0000 UTC m=+1725.499377899 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.490637 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.503605 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzjpc\" (UniqueName: \"kubernetes.io/projected/12be3ae6-8893-4998-94bd-fd12c7c6640c-kube-api-access-hzjpc\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.518829 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.521805 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.554810 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs"] Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.585130 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.585342 4796 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.585381 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzz8g\" (UniqueName: \"kubernetes.io/projected/6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb-kube-api-access-tzz8g\") pod \"rabbitmq-cluster-operator-manager-668c99d594-l2v4r\" (UID: \"6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.585414 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert podName:c2e58505-5b12-41ac-aa51-d857b46b0166 nodeName:}" failed. No retries permitted until 2025-11-27 11:53:28.585397604 +0000 UTC m=+1726.103716522 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" (UID: "c2e58505-5b12-41ac-aa51-d857b46b0166") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.624665 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzz8g\" (UniqueName: \"kubernetes.io/projected/6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb-kube-api-access-tzz8g\") pod \"rabbitmq-cluster-operator-manager-668c99d594-l2v4r\" (UID: \"6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.714222 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-78jwg"] Nov 27 11:53:27 crc kubenswrapper[4796]: W1127 11:53:27.825019 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefe899c4_0891_480f_9aaa_f8b795706142.slice/crio-cf9f5d6d3b2cbb711a918ed698c739788a8fc06c815a30ac53cf90fcaf9db44c WatchSource:0}: Error finding container cf9f5d6d3b2cbb711a918ed698c739788a8fc06c815a30ac53cf90fcaf9db44c: Status 404 returned error can't find the container with id cf9f5d6d3b2cbb711a918ed698c739788a8fc06c815a30ac53cf90fcaf9db44c Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.838091 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" event={"ID":"154fc49c-3eca-4d26-836f-ed87d2d78716","Type":"ContainerStarted","Data":"ffade12b5b9dd46c3dfeec7ce6408c468646487425aa32ccd9b1d0f35ce7b15b"} Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.850594 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" event={"ID":"ca4033f3-d225-4700-bf24-875462a1a404","Type":"ContainerStarted","Data":"b0b4b89bb664cdc0108f5673d5d6c0d20f140fb8d12ba1f0785b5bfcd9bc1aeb"} Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.852071 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" event={"ID":"0c41793c-bff2-4a6a-877c-489df83fa578","Type":"ContainerStarted","Data":"7d075e81ca30103eb9ea1ba3aa13c3993e57d5116cbfe77affde137295ade878"} Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.917227 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.997184 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: I1127 11:53:27.997290 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.997481 4796 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.997550 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:28.99753114 +0000 UTC m=+1726.515850058 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "webhook-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.997575 4796 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 11:53:27 crc kubenswrapper[4796]: E1127 11:53:27.997610 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:28.997600512 +0000 UTC m=+1726.515919430 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "metrics-server-cert" not found Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.018243 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.030501 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.036988 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w"] Nov 27 11:53:28 crc kubenswrapper[4796]: W1127 11:53:28.040366 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5865e45d_8c50_45ea_9fb5_f8a14385de1a.slice/crio-4f8d867576fa7671ddeff52186559142363bb2c727577b28272dee4ffcbde5bb WatchSource:0}: Error finding container 4f8d867576fa7671ddeff52186559142363bb2c727577b28272dee4ffcbde5bb: Status 404 returned error can't find the container with id 4f8d867576fa7671ddeff52186559142363bb2c727577b28272dee4ffcbde5bb Nov 27 11:53:28 crc kubenswrapper[4796]: W1127 11:53:28.042220 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b095ec7_98b1_4e3c_89ed_0bec53da6057.slice/crio-ffe94c638481ba94a029814b09b9435f31819921e6162d66c1f29a0398905b05 WatchSource:0}: Error finding container ffe94c638481ba94a029814b09b9435f31819921e6162d66c1f29a0398905b05: Status 404 returned error can't find the container with id ffe94c638481ba94a029814b09b9435f31819921e6162d66c1f29a0398905b05 Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.422784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.422979 4796 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.423026 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert podName:940eec7d-4d26-462b-958b-ec782cf28fdf nodeName:}" failed. No retries permitted until 2025-11-27 11:53:30.423011013 +0000 UTC m=+1727.941329931 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert") pod "infra-operator-controller-manager-57548d458d-zzhd9" (UID: "940eec7d-4d26-462b-958b-ec782cf28fdf") : secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.434452 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.452740 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.458661 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522"] Nov 27 11:53:28 crc kubenswrapper[4796]: W1127 11:53:28.462888 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6858cdee_21cc_4595_905a_3cd4ce0da27c.slice/crio-b1568e0deee8a5393e5388ec7d458d96869715682adbf3a5b53ee2199018cc15 WatchSource:0}: Error finding container b1568e0deee8a5393e5388ec7d458d96869715682adbf3a5b53ee2199018cc15: Status 404 returned error can't find the container with id b1568e0deee8a5393e5388ec7d458d96869715682adbf3a5b53ee2199018cc15 Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.481920 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.492481 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.498306 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w"] Nov 27 11:53:28 crc kubenswrapper[4796]: W1127 11:53:28.503733 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4905ae06_8f67_4beb_93ca_bcd9e1dca63e.slice/crio-726f6a57899fc85386c0152add7ca3edff9061aaf4c56a661327be79c0a2f989 WatchSource:0}: Error finding container 726f6a57899fc85386c0152add7ca3edff9061aaf4c56a661327be79c0a2f989: Status 404 returned error can't find the container with id 726f6a57899fc85386c0152add7ca3edff9061aaf4c56a661327be79c0a2f989 Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.505468 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.515395 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.522076 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.528166 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm"] Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.534652 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5p44l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-6ks4r_openstack-operators(2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.541575 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5p44l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-6ks4r_openstack-operators(2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.543021 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" podUID="2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.544322 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsbbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-d8k66_openstack-operators(240033a0-6f60-4c73-b2cb-8d4ddb91a66d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.546750 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsbbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-d8k66_openstack-operators(240033a0-6f60-4c73-b2cb-8d4ddb91a66d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.547956 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" podUID="240033a0-6f60-4c73-b2cb-8d4ddb91a66d" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.552009 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp"] Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.560101 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-grznn"] Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.564017 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l5t6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-fsmqp_openstack-operators(d8beead6-b3d4-4f4c-baae-4eaf00fe812a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.566985 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r"] Nov 27 11:53:28 crc kubenswrapper[4796]: W1127 11:53:28.567795 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb72bebe8_8cf0_4e6e_908b_a5095a7a9348.slice/crio-25472e9eec9a7532df0d135c1af3d8dd8cd90b4b20e2491d2cdade9ea1ea9788 WatchSource:0}: Error finding container 25472e9eec9a7532df0d135c1af3d8dd8cd90b4b20e2491d2cdade9ea1ea9788: Status 404 returned error can't find the container with id 25472e9eec9a7532df0d135c1af3d8dd8cd90b4b20e2491d2cdade9ea1ea9788 Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.568042 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l5t6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-fsmqp_openstack-operators(d8beead6-b3d4-4f4c-baae-4eaf00fe812a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.569488 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" podUID="d8beead6-b3d4-4f4c-baae-4eaf00fe812a" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.580633 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-md7p6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-grznn_openstack-operators(b72bebe8-8cf0-4e6e-908b-a5095a7a9348): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.580813 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tzz8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-l2v4r_openstack-operators(6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.581909 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" podUID="6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.584591 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-md7p6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-grznn_openstack-operators(b72bebe8-8cf0-4e6e-908b-a5095a7a9348): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.585769 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" podUID="b72bebe8-8cf0-4e6e-908b-a5095a7a9348" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.624739 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.625950 4796 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.626007 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert podName:c2e58505-5b12-41ac-aa51-d857b46b0166 nodeName:}" failed. No retries permitted until 2025-11-27 11:53:30.625989071 +0000 UTC m=+1728.144307989 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" (UID: "c2e58505-5b12-41ac-aa51-d857b46b0166") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.875407 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" event={"ID":"4905ae06-8f67-4beb-93ca-bcd9e1dca63e","Type":"ContainerStarted","Data":"726f6a57899fc85386c0152add7ca3edff9061aaf4c56a661327be79c0a2f989"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.894534 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" event={"ID":"24bc27f0-611d-4147-9819-6ec0eb012d81","Type":"ContainerStarted","Data":"764a063672cb35233012e8d574146a435c731e7fa75e6ce22b73d390031a44c2"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.912631 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" event={"ID":"0b1822e4-eff9-4ae6-9d91-46bbef3d090b","Type":"ContainerStarted","Data":"b157e0d013488cc7de6dac072ef11252c9522b33cbb69fe8ea1f1de5b3617f95"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.925642 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" event={"ID":"6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb","Type":"ContainerStarted","Data":"41430c80040a4b4cac7a8a85e56ff2ebac297a4d1f5fda69111afddd5d4dc13b"} Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.937431 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" podUID="6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.960139 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" event={"ID":"5865e45d-8c50-45ea-9fb5-f8a14385de1a","Type":"ContainerStarted","Data":"4f8d867576fa7671ddeff52186559142363bb2c727577b28272dee4ffcbde5bb"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.961664 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" event={"ID":"b72bebe8-8cf0-4e6e-908b-a5095a7a9348","Type":"ContainerStarted","Data":"25472e9eec9a7532df0d135c1af3d8dd8cd90b4b20e2491d2cdade9ea1ea9788"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.968848 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" event={"ID":"068c68ca-9efb-4a10-b647-9688a007ddf2","Type":"ContainerStarted","Data":"8e8aa0ba673c78c0f8cf47c8099923de73ec852fd928376000df085d6209708a"} Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.968874 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" podUID="b72bebe8-8cf0-4e6e-908b-a5095a7a9348" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.969770 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" event={"ID":"240033a0-6f60-4c73-b2cb-8d4ddb91a66d","Type":"ContainerStarted","Data":"829eada38a4e07ed96911cd6f983e7ade3782f8cda80d06204cf96e1b780e7b9"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.972859 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" event={"ID":"b0ba2f52-f61b-47b5-8739-3c248d6e012b","Type":"ContainerStarted","Data":"20e313a1af9e14ffbaa2ff3c96d6631a34caf71d3ceae329e4e5363c9988a84d"} Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.973676 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" podUID="240033a0-6f60-4c73-b2cb-8d4ddb91a66d" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.974757 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" event={"ID":"2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1","Type":"ContainerStarted","Data":"63aced0f7d5c23f1d2e414b93ef76bf71e57ed5374ef13fad9a10f9d89705657"} Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.976135 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" event={"ID":"d8beead6-b3d4-4f4c-baae-4eaf00fe812a","Type":"ContainerStarted","Data":"c58c2fb502d26b57fcc009d84bff6d24238e3a72c6dac3b17f94ac8a5bbd35ee"} Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.990991 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" podUID="d8beead6-b3d4-4f4c-baae-4eaf00fe812a" Nov 27 11:53:28 crc kubenswrapper[4796]: E1127 11:53:28.991146 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" podUID="2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1" Nov 27 11:53:28 crc kubenswrapper[4796]: I1127 11:53:28.991164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" event={"ID":"efe899c4-0891-480f-9aaa-f8b795706142","Type":"ContainerStarted","Data":"cf9f5d6d3b2cbb711a918ed698c739788a8fc06c815a30ac53cf90fcaf9db44c"} Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.004384 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" event={"ID":"2b095ec7-98b1-4e3c-89ed-0bec53da6057","Type":"ContainerStarted","Data":"ffe94c638481ba94a029814b09b9435f31819921e6162d66c1f29a0398905b05"} Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.007478 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" event={"ID":"3189b882-80fe-4198-872d-b8b36f0cc766","Type":"ContainerStarted","Data":"303a1f11719e5f2952641180dcd680f9719875e0f1a4fe4daf5447ca37dbe211"} Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.008504 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" event={"ID":"25fd5653-6347-40b2-9d4e-531b58a7e6a3","Type":"ContainerStarted","Data":"dd2c25ab018d38ae5a92f19956d5d38433a2570f4191b3f68d0b7bf046e677d6"} Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.009689 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" event={"ID":"6858cdee-21cc-4595-905a-3cd4ce0da27c","Type":"ContainerStarted","Data":"b1568e0deee8a5393e5388ec7d458d96869715682adbf3a5b53ee2199018cc15"} Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.029045 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" event={"ID":"257e6c0f-c3fd-458e-839a-d28ca580fd13","Type":"ContainerStarted","Data":"726e5747a20d965bae055469e293afcf456a3db9f8e9d6f8f3794fef5835f76a"} Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.059396 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:29 crc kubenswrapper[4796]: I1127 11:53:29.059743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:29 crc kubenswrapper[4796]: E1127 11:53:29.062688 4796 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 11:53:29 crc kubenswrapper[4796]: E1127 11:53:29.063025 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:31.06300506 +0000 UTC m=+1728.581323988 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "metrics-server-cert" not found Nov 27 11:53:29 crc kubenswrapper[4796]: E1127 11:53:29.064239 4796 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 11:53:29 crc kubenswrapper[4796]: E1127 11:53:29.064315 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:31.064299476 +0000 UTC m=+1728.582618474 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "webhook-server-cert" not found Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.055163 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" podUID="d8beead6-b3d4-4f4c-baae-4eaf00fe812a" Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.055163 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" podUID="240033a0-6f60-4c73-b2cb-8d4ddb91a66d" Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.057104 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" podUID="2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1" Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.058670 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" podUID="b72bebe8-8cf0-4e6e-908b-a5095a7a9348" Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.067623 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" podUID="6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb" Nov 27 11:53:30 crc kubenswrapper[4796]: I1127 11:53:30.503327 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.503571 4796 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.503681 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert podName:940eec7d-4d26-462b-958b-ec782cf28fdf nodeName:}" failed. No retries permitted until 2025-11-27 11:53:34.503659756 +0000 UTC m=+1732.021978674 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert") pod "infra-operator-controller-manager-57548d458d-zzhd9" (UID: "940eec7d-4d26-462b-958b-ec782cf28fdf") : secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:30 crc kubenswrapper[4796]: I1127 11:53:30.705948 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.706116 4796 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:30 crc kubenswrapper[4796]: E1127 11:53:30.706217 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert podName:c2e58505-5b12-41ac-aa51-d857b46b0166 nodeName:}" failed. No retries permitted until 2025-11-27 11:53:34.706190293 +0000 UTC m=+1732.224509211 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" (UID: "c2e58505-5b12-41ac-aa51-d857b46b0166") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.111347 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.111493 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:31 crc kubenswrapper[4796]: E1127 11:53:31.111699 4796 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 11:53:31 crc kubenswrapper[4796]: E1127 11:53:31.111761 4796 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 11:53:31 crc kubenswrapper[4796]: E1127 11:53:31.111768 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:35.11175011 +0000 UTC m=+1732.630069028 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "metrics-server-cert" not found Nov 27 11:53:31 crc kubenswrapper[4796]: E1127 11:53:31.111866 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:35.111799392 +0000 UTC m=+1732.630118310 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "webhook-server-cert" not found Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.879528 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.879604 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.879776 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.880357 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:53:31 crc kubenswrapper[4796]: I1127 11:53:31.880433 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" gracePeriod=600 Nov 27 11:53:32 crc kubenswrapper[4796]: I1127 11:53:32.075692 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" exitCode=0 Nov 27 11:53:32 crc kubenswrapper[4796]: I1127 11:53:32.076051 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8"} Nov 27 11:53:32 crc kubenswrapper[4796]: I1127 11:53:32.076206 4796 scope.go:117] "RemoveContainer" containerID="8b7cb903692ced3867395064a71bd00da73fa8daa661bb81331f102ad27ef1c9" Nov 27 11:53:34 crc kubenswrapper[4796]: I1127 11:53:34.566675 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:34 crc kubenswrapper[4796]: E1127 11:53:34.566842 4796 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:34 crc kubenswrapper[4796]: E1127 11:53:34.567142 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert podName:940eec7d-4d26-462b-958b-ec782cf28fdf nodeName:}" failed. No retries permitted until 2025-11-27 11:53:42.567120875 +0000 UTC m=+1740.085439793 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert") pod "infra-operator-controller-manager-57548d458d-zzhd9" (UID: "940eec7d-4d26-462b-958b-ec782cf28fdf") : secret "infra-operator-webhook-server-cert" not found Nov 27 11:53:34 crc kubenswrapper[4796]: I1127 11:53:34.769901 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:34 crc kubenswrapper[4796]: E1127 11:53:34.770082 4796 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:34 crc kubenswrapper[4796]: E1127 11:53:34.770150 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert podName:c2e58505-5b12-41ac-aa51-d857b46b0166 nodeName:}" failed. No retries permitted until 2025-11-27 11:53:42.770129275 +0000 UTC m=+1740.288448193 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" (UID: "c2e58505-5b12-41ac-aa51-d857b46b0166") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 11:53:35 crc kubenswrapper[4796]: I1127 11:53:35.177638 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:35 crc kubenswrapper[4796]: I1127 11:53:35.177819 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:35 crc kubenswrapper[4796]: E1127 11:53:35.178247 4796 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 11:53:35 crc kubenswrapper[4796]: E1127 11:53:35.178247 4796 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 11:53:35 crc kubenswrapper[4796]: E1127 11:53:35.178341 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:43.178322064 +0000 UTC m=+1740.696640982 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "webhook-server-cert" not found Nov 27 11:53:35 crc kubenswrapper[4796]: E1127 11:53:35.178388 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs podName:12be3ae6-8893-4998-94bd-fd12c7c6640c nodeName:}" failed. No retries permitted until 2025-11-27 11:53:43.178363645 +0000 UTC m=+1740.696682583 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs") pod "openstack-operator-controller-manager-557965c5b6-9qjgg" (UID: "12be3ae6-8893-4998-94bd-fd12c7c6640c") : secret "metrics-server-cert" not found Nov 27 11:53:40 crc kubenswrapper[4796]: E1127 11:53:40.751106 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:53:40 crc kubenswrapper[4796]: E1127 11:53:40.900867 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:5cfdc7db2f744471a51fa07c9e3c28e15be2d043" Nov 27 11:53:40 crc kubenswrapper[4796]: E1127 11:53:40.900937 4796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:5cfdc7db2f744471a51fa07c9e3c28e15be2d043" Nov 27 11:53:40 crc kubenswrapper[4796]: E1127 11:53:40.901135 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.20:5001/openstack-k8s-operators/telemetry-operator:5cfdc7db2f744471a51fa07c9e3c28e15be2d043,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wrvzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6f77465788-gr9vn_openstack-operators(068c68ca-9efb-4a10-b647-9688a007ddf2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:53:41 crc kubenswrapper[4796]: I1127 11:53:41.141030 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:53:41 crc kubenswrapper[4796]: E1127 11:53:41.141562 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:53:41 crc kubenswrapper[4796]: E1127 11:53:41.461490 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711" Nov 27 11:53:41 crc kubenswrapper[4796]: E1127 11:53:41.461942 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x2jwn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-drlr5_openstack-operators(5865e45d-8c50-45ea-9fb5-f8a14385de1a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:53:42 crc kubenswrapper[4796]: E1127 11:53:42.115476 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jb9tw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-4xs4w_openstack-operators(b0ba2f52-f61b-47b5-8739-3c248d6e012b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:42 crc kubenswrapper[4796]: E1127 11:53:42.117190 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" podUID="b0ba2f52-f61b-47b5-8739-3c248d6e012b" Nov 27 11:53:42 crc kubenswrapper[4796]: E1127 11:53:42.117341 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kjgf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7b64f4fb85-z5fwq_openstack-operators(ca4033f3-d225-4700-bf24-875462a1a404): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 11:53:42 crc kubenswrapper[4796]: E1127 11:53:42.118663 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" podUID="ca4033f3-d225-4700-bf24-875462a1a404" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.156508 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" event={"ID":"2b095ec7-98b1-4e3c-89ed-0bec53da6057","Type":"ContainerStarted","Data":"2e5d56798465928953b3015d0ac7434d29a730d3c2504c3c6fc8488749650259"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.159203 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" event={"ID":"b0ba2f52-f61b-47b5-8739-3c248d6e012b","Type":"ContainerStarted","Data":"c2ccc2e6c08d333a536ab15c87cf46d8d0915563c0c7b9b5a6ab3addecc366d1"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.159880 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:42 crc kubenswrapper[4796]: E1127 11:53:42.162466 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" podUID="b0ba2f52-f61b-47b5-8739-3c248d6e012b" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.189617 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" event={"ID":"0c41793c-bff2-4a6a-877c-489df83fa578","Type":"ContainerStarted","Data":"acf5fc11fe47497c33d6641f133fb3ea91aaee3141f06a70ab71273c60ea5bea"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.227642 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" event={"ID":"257e6c0f-c3fd-458e-839a-d28ca580fd13","Type":"ContainerStarted","Data":"c409507c2135f5f27570e39bd68e6779cba65d32edb6784d94784048405e9a7d"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.241493 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" event={"ID":"0b1822e4-eff9-4ae6-9d91-46bbef3d090b","Type":"ContainerStarted","Data":"32678e2625a53931e333be55ec86d72927e971bef4a924a36cb601bdf4662d4e"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.245686 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" event={"ID":"efe899c4-0891-480f-9aaa-f8b795706142","Type":"ContainerStarted","Data":"9ed917eb1c33d14ca975b7bce7f8941b6a46e9a20fcdadd27c6bab2bc422b505"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.246855 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" event={"ID":"ca4033f3-d225-4700-bf24-875462a1a404","Type":"ContainerStarted","Data":"b44cb83885716804b9c59246da41da1105a395191295a8bb5e3586aeef71650d"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.247249 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:42 crc kubenswrapper[4796]: E1127 11:53:42.249164 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" podUID="ca4033f3-d225-4700-bf24-875462a1a404" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.261690 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" event={"ID":"24bc27f0-611d-4147-9819-6ec0eb012d81","Type":"ContainerStarted","Data":"0497e1e8d3cef226778ec265d805d478dea6678db775f207743a3e6e8c2ed080"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.285390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" event={"ID":"4905ae06-8f67-4beb-93ca-bcd9e1dca63e","Type":"ContainerStarted","Data":"ce4669a8f5cf27dde35df8fbed376b7bd4ff69fe2f26d3b77c3fe05194fb9bb2"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.300400 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" event={"ID":"3189b882-80fe-4198-872d-b8b36f0cc766","Type":"ContainerStarted","Data":"5ac952c11da3e95815f4e0fbbe9032e617f0ed4887416d06603327d73329694a"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.309448 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" event={"ID":"25fd5653-6347-40b2-9d4e-531b58a7e6a3","Type":"ContainerStarted","Data":"0521cdb891cecc5a147f014c3e184d39320d52801f03eb3a18513391008742dc"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.325390 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" event={"ID":"154fc49c-3eca-4d26-836f-ed87d2d78716","Type":"ContainerStarted","Data":"1161ac5dc1003576d1f9c0859719a8e4fcfe4c37c06a456ee53a3301214212be"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.340408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" event={"ID":"6858cdee-21cc-4595-905a-3cd4ce0da27c","Type":"ContainerStarted","Data":"66dda1830910c9005dca2279f54b7dd486a4bf58f96d6f709fcefd9e7c435670"} Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.611829 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.621105 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/940eec7d-4d26-462b-958b-ec782cf28fdf-cert\") pod \"infra-operator-controller-manager-57548d458d-zzhd9\" (UID: \"940eec7d-4d26-462b-958b-ec782cf28fdf\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.674529 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.815350 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.823925 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e58505-5b12-41ac-aa51-d857b46b0166-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w\" (UID: \"c2e58505-5b12-41ac-aa51-d857b46b0166\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:42 crc kubenswrapper[4796]: I1127 11:53:42.904171 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:53:43 crc kubenswrapper[4796]: I1127 11:53:43.221345 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:43 crc kubenswrapper[4796]: I1127 11:53:43.221506 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:43 crc kubenswrapper[4796]: I1127 11:53:43.229998 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-webhook-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:43 crc kubenswrapper[4796]: I1127 11:53:43.234074 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12be3ae6-8893-4998-94bd-fd12c7c6640c-metrics-certs\") pod \"openstack-operator-controller-manager-557965c5b6-9qjgg\" (UID: \"12be3ae6-8893-4998-94bd-fd12c7c6640c\") " pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:43 crc kubenswrapper[4796]: E1127 11:53:43.350248 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" podUID="ca4033f3-d225-4700-bf24-875462a1a404" Nov 27 11:53:43 crc kubenswrapper[4796]: E1127 11:53:43.350358 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" podUID="b0ba2f52-f61b-47b5-8739-3c248d6e012b" Nov 27 11:53:43 crc kubenswrapper[4796]: I1127 11:53:43.478605 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:43 crc kubenswrapper[4796]: I1127 11:53:43.826091 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9"] Nov 27 11:53:44 crc kubenswrapper[4796]: I1127 11:53:44.006712 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w"] Nov 27 11:53:44 crc kubenswrapper[4796]: W1127 11:53:44.524011 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2e58505_5b12_41ac_aa51_d857b46b0166.slice/crio-8ea81a335cf909303056ef477a5bab5cdd7a03face16bdc383cef4ed603153cd WatchSource:0}: Error finding container 8ea81a335cf909303056ef477a5bab5cdd7a03face16bdc383cef4ed603153cd: Status 404 returned error can't find the container with id 8ea81a335cf909303056ef477a5bab5cdd7a03face16bdc383cef4ed603153cd Nov 27 11:53:44 crc kubenswrapper[4796]: W1127 11:53:44.530949 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod940eec7d_4d26_462b_958b_ec782cf28fdf.slice/crio-9d7ffd06a9501376371737e8d0eb7fa14ea2ae4f77185eedf06109752d05a231 WatchSource:0}: Error finding container 9d7ffd06a9501376371737e8d0eb7fa14ea2ae4f77185eedf06109752d05a231: Status 404 returned error can't find the container with id 9d7ffd06a9501376371737e8d0eb7fa14ea2ae4f77185eedf06109752d05a231 Nov 27 11:53:45 crc kubenswrapper[4796]: I1127 11:53:45.368890 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" event={"ID":"c2e58505-5b12-41ac-aa51-d857b46b0166","Type":"ContainerStarted","Data":"8ea81a335cf909303056ef477a5bab5cdd7a03face16bdc383cef4ed603153cd"} Nov 27 11:53:45 crc kubenswrapper[4796]: I1127 11:53:45.370117 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" event={"ID":"940eec7d-4d26-462b-958b-ec782cf28fdf","Type":"ContainerStarted","Data":"9d7ffd06a9501376371737e8d0eb7fa14ea2ae4f77185eedf06109752d05a231"} Nov 27 11:53:45 crc kubenswrapper[4796]: I1127 11:53:45.830805 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg"] Nov 27 11:53:46 crc kubenswrapper[4796]: W1127 11:53:46.720409 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12be3ae6_8893_4998_94bd_fd12c7c6640c.slice/crio-6460f08ed7741861b1a40469f2554605386dfe8d72dcb8dc254ee787f20fce8d WatchSource:0}: Error finding container 6460f08ed7741861b1a40469f2554605386dfe8d72dcb8dc254ee787f20fce8d: Status 404 returned error can't find the container with id 6460f08ed7741861b1a40469f2554605386dfe8d72dcb8dc254ee787f20fce8d Nov 27 11:53:47 crc kubenswrapper[4796]: I1127 11:53:47.324173 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" Nov 27 11:53:47 crc kubenswrapper[4796]: E1127 11:53:47.325824 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" podUID="b0ba2f52-f61b-47b5-8739-3c248d6e012b" Nov 27 11:53:47 crc kubenswrapper[4796]: I1127 11:53:47.399967 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" event={"ID":"12be3ae6-8893-4998-94bd-fd12c7c6640c","Type":"ContainerStarted","Data":"6460f08ed7741861b1a40469f2554605386dfe8d72dcb8dc254ee787f20fce8d"} Nov 27 11:53:52 crc kubenswrapper[4796]: I1127 11:53:52.569363 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:53:52 crc kubenswrapper[4796]: E1127 11:53:52.570525 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:53:53 crc kubenswrapper[4796]: I1127 11:53:53.460116 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" event={"ID":"12be3ae6-8893-4998-94bd-fd12c7c6640c","Type":"ContainerStarted","Data":"0343e8b18c5eb977366c68920982c254452a6641412af95e692bfa35a3b69ff0"} Nov 27 11:53:53 crc kubenswrapper[4796]: I1127 11:53:53.460419 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:53:53 crc kubenswrapper[4796]: I1127 11:53:53.499537 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" podStartSLOduration=26.499508722 podStartE2EDuration="26.499508722s" podCreationTimestamp="2025-11-27 11:53:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:53:53.489651019 +0000 UTC m=+1751.007970017" watchObservedRunningTime="2025-11-27 11:53:53.499508722 +0000 UTC m=+1751.017827670" Nov 27 11:53:56 crc kubenswrapper[4796]: I1127 11:53:56.571291 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" Nov 27 11:53:59 crc kubenswrapper[4796]: I1127 11:53:59.538080 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" event={"ID":"240033a0-6f60-4c73-b2cb-8d4ddb91a66d","Type":"ContainerStarted","Data":"e6888761b54e9d619189173836154ce682257cf214cb641d94f5bcc70bbbe367"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.546685 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" event={"ID":"0c41793c-bff2-4a6a-877c-489df83fa578","Type":"ContainerStarted","Data":"1f42cd4b4049a4f1687913a9292353475b7a64a53040610eb8b02c8e3b575411"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.548282 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" event={"ID":"0b1822e4-eff9-4ae6-9d91-46bbef3d090b","Type":"ContainerStarted","Data":"b6b5c201b039c903d02d8878256e4e54c934b87df2f110fe07c449e12cc3c3e5"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.549149 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" event={"ID":"940eec7d-4d26-462b-958b-ec782cf28fdf","Type":"ContainerStarted","Data":"df3ca21b2cdd1c037cc40abd633918f26103708f7bb5073451b7cf03dc07306f"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.550648 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" event={"ID":"154fc49c-3eca-4d26-836f-ed87d2d78716","Type":"ContainerStarted","Data":"b03d2a6e4194a43b04b4b23e2ebcf4ffbbe74f96df604ef28929fc47e20307bf"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.552170 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" event={"ID":"5865e45d-8c50-45ea-9fb5-f8a14385de1a","Type":"ContainerStarted","Data":"1536ca90a38ce246b9f941b457276cca640379a6e059e73763740b7bf9adec1a"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.553994 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" event={"ID":"b72bebe8-8cf0-4e6e-908b-a5095a7a9348","Type":"ContainerStarted","Data":"2bbb6edb394be61966b80d524872830fab28fb0942451855aaa28b09d81cbfe4"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.555225 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" event={"ID":"c2e58505-5b12-41ac-aa51-d857b46b0166","Type":"ContainerStarted","Data":"99f1783eee0c58a0eeed6c7354fd4fad9d89008738709549a28afc84f25743df"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.556937 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" event={"ID":"068c68ca-9efb-4a10-b647-9688a007ddf2","Type":"ContainerStarted","Data":"94ef9f3de05fb57d2161e635b4d599477ece34a6cfaa8e7dc998c477ec9d38de"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.558661 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" event={"ID":"25fd5653-6347-40b2-9d4e-531b58a7e6a3","Type":"ContainerStarted","Data":"3923741dfacb44eb100d32f2a6f8af14c89c7488424df6416e3ae231f46a8f6e"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.560396 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" event={"ID":"2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1","Type":"ContainerStarted","Data":"4db1d0b236b892d9112487285a20b07d3af8e36c768132fabdd3cd3b79fb6c0e"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.562700 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" event={"ID":"257e6c0f-c3fd-458e-839a-d28ca580fd13","Type":"ContainerStarted","Data":"2a57c422b268b1b16408d4ad8f2efc8408c35c17883f2da6f92f7ba169b11970"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.563956 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" event={"ID":"d8beead6-b3d4-4f4c-baae-4eaf00fe812a","Type":"ContainerStarted","Data":"60532fd7d6f661b79e09896e4c7ccb67b441d4532b3a3b49b396618756415454"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.565697 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" event={"ID":"2b095ec7-98b1-4e3c-89ed-0bec53da6057","Type":"ContainerStarted","Data":"c8530a955a1f5802b95bb29705489241c82d34ee700c50db2f063fe35dff2b40"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.572845 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" event={"ID":"3189b882-80fe-4198-872d-b8b36f0cc766","Type":"ContainerStarted","Data":"611c39c508dcef2f7ee685887d14f607b603b80465c0910f96fcec7472beaedb"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.574481 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" event={"ID":"6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb","Type":"ContainerStarted","Data":"15ec24c3bd57564fd73dec28f9852356ba7b4a9fe569efd32a7b35e6db8db792"} Nov 27 11:54:00 crc kubenswrapper[4796]: I1127 11:54:00.592598 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" event={"ID":"4905ae06-8f67-4beb-93ca-bcd9e1dca63e","Type":"ContainerStarted","Data":"063ef462336f66ac67fe63c9c9f7bb94d1d5ff9c0006c635961d229eadb3eda9"} Nov 27 11:54:01 crc kubenswrapper[4796]: E1127 11:54:01.429670 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" podUID="068c68ca-9efb-4a10-b647-9688a007ddf2" Nov 27 11:54:01 crc kubenswrapper[4796]: E1127 11:54:01.430178 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" podUID="5865e45d-8c50-45ea-9fb5-f8a14385de1a" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.601556 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" event={"ID":"b72bebe8-8cf0-4e6e-908b-a5095a7a9348","Type":"ContainerStarted","Data":"61b8b8143234742e72d63eadc4bc05d3df45da57b9ac8c5ac940f0dc667ad2af"} Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.611455 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" event={"ID":"240033a0-6f60-4c73-b2cb-8d4ddb91a66d","Type":"ContainerStarted","Data":"3ad38f7820198d4f28fc18a425fec7f0d6f1c3061022fb41fa79a01ef2eca1d2"} Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.611508 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.613035 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.613053 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.613063 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.613073 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.615528 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.615866 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.615901 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.615918 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.615928 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.616897 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.617936 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.619058 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.621809 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.622090 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.622587 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.641174 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-88bxr" podStartSLOduration=11.536700576 podStartE2EDuration="35.641157017s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.509176748 +0000 UTC m=+1726.027495666" lastFinishedPulling="2025-11-27 11:53:52.613633189 +0000 UTC m=+1750.131952107" observedRunningTime="2025-11-27 11:54:01.63866922 +0000 UTC m=+1759.156988168" watchObservedRunningTime="2025-11-27 11:54:01.641157017 +0000 UTC m=+1759.159475935" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.659881 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcrbs" podStartSLOduration=10.76708249 podStartE2EDuration="35.659858475s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:27.702873475 +0000 UTC m=+1725.221192393" lastFinishedPulling="2025-11-27 11:53:52.59564946 +0000 UTC m=+1750.113968378" observedRunningTime="2025-11-27 11:54:01.657955945 +0000 UTC m=+1759.176274873" watchObservedRunningTime="2025-11-27 11:54:01.659858475 +0000 UTC m=+1759.178177403" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.712628 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-cjtbm" podStartSLOduration=11.682626724 podStartE2EDuration="35.712605592s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.533853743 +0000 UTC m=+1726.052172661" lastFinishedPulling="2025-11-27 11:53:52.563832611 +0000 UTC m=+1750.082151529" observedRunningTime="2025-11-27 11:54:01.702722109 +0000 UTC m=+1759.221041057" watchObservedRunningTime="2025-11-27 11:54:01.712605592 +0000 UTC m=+1759.230924520" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.751028 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-6fpn9" podStartSLOduration=11.067507574 podStartE2EDuration="35.751007277s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.019025018 +0000 UTC m=+1725.537343936" lastFinishedPulling="2025-11-27 11:53:52.702524721 +0000 UTC m=+1750.220843639" observedRunningTime="2025-11-27 11:54:01.747819492 +0000 UTC m=+1759.266138450" watchObservedRunningTime="2025-11-27 11:54:01.751007277 +0000 UTC m=+1759.269326215" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.771755 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-l2v4r" podStartSLOduration=11.536630347 podStartE2EDuration="34.77173524s" podCreationTimestamp="2025-11-27 11:53:27 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.580708703 +0000 UTC m=+1726.099027621" lastFinishedPulling="2025-11-27 11:53:51.815813556 +0000 UTC m=+1749.334132514" observedRunningTime="2025-11-27 11:54:01.766949453 +0000 UTC m=+1759.285268371" watchObservedRunningTime="2025-11-27 11:54:01.77173524 +0000 UTC m=+1759.290054148" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.785394 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-8lq5v" podStartSLOduration=11.624467028 podStartE2EDuration="35.785377684s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.482154169 +0000 UTC m=+1726.000473087" lastFinishedPulling="2025-11-27 11:53:52.643064825 +0000 UTC m=+1750.161383743" observedRunningTime="2025-11-27 11:54:01.784632764 +0000 UTC m=+1759.302951682" watchObservedRunningTime="2025-11-27 11:54:01.785377684 +0000 UTC m=+1759.303696602" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.826952 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-zqg7w" podStartSLOduration=11.283667538 podStartE2EDuration="35.826927362s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.045671236 +0000 UTC m=+1725.563990154" lastFinishedPulling="2025-11-27 11:53:52.58893106 +0000 UTC m=+1750.107249978" observedRunningTime="2025-11-27 11:54:01.819168735 +0000 UTC m=+1759.337487653" watchObservedRunningTime="2025-11-27 11:54:01.826927362 +0000 UTC m=+1759.345246280" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.852667 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-zlbmz" podStartSLOduration=10.910648219 podStartE2EDuration="35.852644148s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:27.700202112 +0000 UTC m=+1725.218521030" lastFinishedPulling="2025-11-27 11:53:52.642198041 +0000 UTC m=+1750.160516959" observedRunningTime="2025-11-27 11:54:01.843021452 +0000 UTC m=+1759.361340370" watchObservedRunningTime="2025-11-27 11:54:01.852644148 +0000 UTC m=+1759.370963066" Nov 27 11:54:01 crc kubenswrapper[4796]: I1127 11:54:01.879861 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-8wjn5" podStartSLOduration=11.871947745 podStartE2EDuration="35.879840394s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.533434311 +0000 UTC m=+1726.051753229" lastFinishedPulling="2025-11-27 11:53:52.54132695 +0000 UTC m=+1750.059645878" observedRunningTime="2025-11-27 11:54:01.867733231 +0000 UTC m=+1759.386052159" watchObservedRunningTime="2025-11-27 11:54:01.879840394 +0000 UTC m=+1759.398159312" Nov 27 11:54:01 crc kubenswrapper[4796]: E1127 11:54:01.889360 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 27 11:54:01 crc kubenswrapper[4796]: E1127 11:54:01.889619 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c94x9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5d494799bf-wkg8j_openstack-operators(24bc27f0-611d-4147-9819-6ec0eb012d81): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:01 crc kubenswrapper[4796]: E1127 11:54:01.895251 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" podUID="24bc27f0-611d-4147-9819-6ec0eb012d81" Nov 27 11:54:02 crc kubenswrapper[4796]: E1127 11:54:02.239020 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 27 11:54:02 crc kubenswrapper[4796]: E1127 11:54:02.239168 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-76ldn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-955677c94-78jwg_openstack-operators(efe899c4-0891-480f-9aaa-f8b795706142): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:02 crc kubenswrapper[4796]: E1127 11:54:02.240335 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" podUID="efe899c4-0891-480f-9aaa-f8b795706142" Nov 27 11:54:02 crc kubenswrapper[4796]: E1127 11:54:02.259077 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 27 11:54:02 crc kubenswrapper[4796]: E1127 11:54:02.259220 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jdngf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5d499bf58b-ll522_openstack-operators(6858cdee-21cc-4595-905a-3cd4ce0da27c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:02 crc kubenswrapper[4796]: E1127 11:54:02.260345 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" podUID="6858cdee-21cc-4595-905a-3cd4ce0da27c" Nov 27 11:54:02 crc kubenswrapper[4796]: I1127 11:54:02.683982 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" podStartSLOduration=14.375376868 podStartE2EDuration="36.683965605s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.580398715 +0000 UTC m=+1726.098717633" lastFinishedPulling="2025-11-27 11:53:50.888987452 +0000 UTC m=+1748.407306370" observedRunningTime="2025-11-27 11:54:02.682795825 +0000 UTC m=+1760.201114763" watchObservedRunningTime="2025-11-27 11:54:02.683965605 +0000 UTC m=+1760.202284533" Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.485068 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-557965c5b6-9qjgg" Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.621612 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" event={"ID":"d8beead6-b3d4-4f4c-baae-4eaf00fe812a","Type":"ContainerStarted","Data":"420e8ee29d4fe4a148fbeeaf2ac1a2fb77e3c10cced93edf017ddedc16740b39"} Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.623149 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" event={"ID":"efe899c4-0891-480f-9aaa-f8b795706142","Type":"ContainerStarted","Data":"d1a209cedd0c9bffdeda0ee7117114211b324220957e4f2f83a6e35508fbc442"} Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.624531 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" event={"ID":"c2e58505-5b12-41ac-aa51-d857b46b0166","Type":"ContainerStarted","Data":"c98cd6a9b72cee1890ed6f7cd96afcce0d5b3f4b3627c9301f1176004bcd88fd"} Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.625762 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" event={"ID":"940eec7d-4d26-462b-958b-ec782cf28fdf","Type":"ContainerStarted","Data":"d3b892d7aa0afc09d55c4af2ebe798b0a21467b9c4d115f6e0d98dae2a0607e9"} Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.627330 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" event={"ID":"2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1","Type":"ContainerStarted","Data":"29ab171f211bfa0d87cd3f20519b53bfd11ec9a993d09922493ed1f5ba1921d3"} Nov 27 11:54:03 crc kubenswrapper[4796]: I1127 11:54:03.628801 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" event={"ID":"24bc27f0-611d-4147-9819-6ec0eb012d81","Type":"ContainerStarted","Data":"5103904e571ccaa05829ae66e7d71187181e09e170666491b56ccccc5b53246a"} Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.640322 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" event={"ID":"6858cdee-21cc-4595-905a-3cd4ce0da27c","Type":"ContainerStarted","Data":"efe9789c65273cf4198cb8f606e4b6c201f32c18e48c106814113db29415d021"} Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.642998 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" event={"ID":"b0ba2f52-f61b-47b5-8739-3c248d6e012b","Type":"ContainerStarted","Data":"9e9623e78d85eae0ca6638f28f5ac692e280cc11306879d1f9e575661dcd5ff0"} Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.644069 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.645990 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.667143 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" podStartSLOduration=15.462563868 podStartE2EDuration="38.6671233s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.544181625 +0000 UTC m=+1726.062500543" lastFinishedPulling="2025-11-27 11:53:51.748741057 +0000 UTC m=+1749.267059975" observedRunningTime="2025-11-27 11:54:03.6441461 +0000 UTC m=+1761.162465028" watchObservedRunningTime="2025-11-27 11:54:04.6671233 +0000 UTC m=+1762.185442218" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.667568 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" podStartSLOduration=25.680575834 podStartE2EDuration="38.667559961s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.457200047 +0000 UTC m=+1725.975518965" lastFinishedPulling="2025-11-27 11:53:41.444184174 +0000 UTC m=+1738.962503092" observedRunningTime="2025-11-27 11:54:04.662086985 +0000 UTC m=+1762.180405913" watchObservedRunningTime="2025-11-27 11:54:04.667559961 +0000 UTC m=+1762.185878879" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.714616 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" podStartSLOduration=15.571212 podStartE2EDuration="38.714594766s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.53447174 +0000 UTC m=+1726.052790658" lastFinishedPulling="2025-11-27 11:53:51.677854506 +0000 UTC m=+1749.196173424" observedRunningTime="2025-11-27 11:54:04.714594266 +0000 UTC m=+1762.232913194" watchObservedRunningTime="2025-11-27 11:54:04.714594766 +0000 UTC m=+1762.232913684" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.718804 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" podStartSLOduration=31.454927805 podStartE2EDuration="38.718791748s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:44.548712597 +0000 UTC m=+1742.067031515" lastFinishedPulling="2025-11-27 11:53:51.81257655 +0000 UTC m=+1749.330895458" observedRunningTime="2025-11-27 11:54:04.69375673 +0000 UTC m=+1762.212075668" watchObservedRunningTime="2025-11-27 11:54:04.718791748 +0000 UTC m=+1762.237110666" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.741825 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" podStartSLOduration=15.557009001 podStartE2EDuration="38.741807283s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.563844002 +0000 UTC m=+1726.082162910" lastFinishedPulling="2025-11-27 11:53:51.748642274 +0000 UTC m=+1749.266961192" observedRunningTime="2025-11-27 11:54:04.738303548 +0000 UTC m=+1762.256622466" watchObservedRunningTime="2025-11-27 11:54:04.741807283 +0000 UTC m=+1762.260126201" Nov 27 11:54:04 crc kubenswrapper[4796]: I1127 11:54:04.768697 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" podStartSLOduration=31.502309309 podStartE2EDuration="38.768677029s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:44.54845272 +0000 UTC m=+1742.066771638" lastFinishedPulling="2025-11-27 11:53:51.81482044 +0000 UTC m=+1749.333139358" observedRunningTime="2025-11-27 11:54:04.766823399 +0000 UTC m=+1762.285142327" watchObservedRunningTime="2025-11-27 11:54:04.768677029 +0000 UTC m=+1762.286995947" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.654605 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.656200 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.666409 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.666691 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.715707 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-4xs4w" podStartSLOduration=26.778690183 podStartE2EDuration="39.715671742s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.507926414 +0000 UTC m=+1726.026245332" lastFinishedPulling="2025-11-27 11:53:41.444907953 +0000 UTC m=+1738.963226891" observedRunningTime="2025-11-27 11:54:05.684527941 +0000 UTC m=+1763.202846929" watchObservedRunningTime="2025-11-27 11:54:05.715671742 +0000 UTC m=+1763.233990700" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.727011 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-ll522" podStartSLOduration=26.742500365 podStartE2EDuration="39.726976713s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.467481058 +0000 UTC m=+1725.985799976" lastFinishedPulling="2025-11-27 11:53:41.451957406 +0000 UTC m=+1738.970276324" observedRunningTime="2025-11-27 11:54:05.714134721 +0000 UTC m=+1763.232453679" watchObservedRunningTime="2025-11-27 11:54:05.726976713 +0000 UTC m=+1763.245295671" Nov 27 11:54:05 crc kubenswrapper[4796]: I1127 11:54:05.763105 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-78jwg" podStartSLOduration=26.154001862 podStartE2EDuration="39.763075836s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:27.835233123 +0000 UTC m=+1725.353552041" lastFinishedPulling="2025-11-27 11:53:41.444307097 +0000 UTC m=+1738.962626015" observedRunningTime="2025-11-27 11:54:05.744958964 +0000 UTC m=+1763.263277892" watchObservedRunningTime="2025-11-27 11:54:05.763075836 +0000 UTC m=+1763.281394804" Nov 27 11:54:06 crc kubenswrapper[4796]: I1127 11:54:06.664084 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" event={"ID":"ca4033f3-d225-4700-bf24-875462a1a404","Type":"ContainerStarted","Data":"b6490cbdc85f63e7ef72710dd5d44c120090a725889c282c5d93639b4240433f"} Nov 27 11:54:06 crc kubenswrapper[4796]: I1127 11:54:06.687939 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-z5fwq" podStartSLOduration=27.348503276 podStartE2EDuration="40.687920188s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:27.517932929 +0000 UTC m=+1725.036251847" lastFinishedPulling="2025-11-27 11:53:40.857349841 +0000 UTC m=+1738.375668759" observedRunningTime="2025-11-27 11:54:06.683976763 +0000 UTC m=+1764.202295681" watchObservedRunningTime="2025-11-27 11:54:06.687920188 +0000 UTC m=+1764.206239106" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.275020 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.277185 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.277507 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-6ks4r" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.286967 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-fsmqp" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.418714 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.421033 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-grznn" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.491591 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.494415 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-d8k66" Nov 27 11:54:07 crc kubenswrapper[4796]: I1127 11:54:07.572353 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:54:07 crc kubenswrapper[4796]: E1127 11:54:07.573127 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:54:09 crc kubenswrapper[4796]: I1127 11:54:09.689684 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" event={"ID":"5865e45d-8c50-45ea-9fb5-f8a14385de1a","Type":"ContainerStarted","Data":"69c077d3ef67d078386d84ea086ea43cead3412fe85aaea0ecbd505626deca6d"} Nov 27 11:54:09 crc kubenswrapper[4796]: I1127 11:54:09.690139 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:54:09 crc kubenswrapper[4796]: I1127 11:54:09.691521 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" event={"ID":"068c68ca-9efb-4a10-b647-9688a007ddf2","Type":"ContainerStarted","Data":"5d9ec877a7ec64fd4839335d75b915d752d56e4ce8dc4bd270757f829c5fcd33"} Nov 27 11:54:09 crc kubenswrapper[4796]: I1127 11:54:09.691677 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:54:09 crc kubenswrapper[4796]: I1127 11:54:09.707819 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" podStartSLOduration=3.249621581 podStartE2EDuration="43.707799877s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.042115239 +0000 UTC m=+1725.560434157" lastFinishedPulling="2025-11-27 11:54:08.500293535 +0000 UTC m=+1766.018612453" observedRunningTime="2025-11-27 11:54:09.707623842 +0000 UTC m=+1767.225942760" watchObservedRunningTime="2025-11-27 11:54:09.707799877 +0000 UTC m=+1767.226118805" Nov 27 11:54:09 crc kubenswrapper[4796]: I1127 11:54:09.726474 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" podStartSLOduration=3.762000296 podStartE2EDuration="43.726447985s" podCreationTimestamp="2025-11-27 11:53:26 +0000 UTC" firstStartedPulling="2025-11-27 11:53:28.534148291 +0000 UTC m=+1726.052467209" lastFinishedPulling="2025-11-27 11:54:08.49859597 +0000 UTC m=+1766.016914898" observedRunningTime="2025-11-27 11:54:09.725726496 +0000 UTC m=+1767.244045424" watchObservedRunningTime="2025-11-27 11:54:09.726447985 +0000 UTC m=+1767.244766913" Nov 27 11:54:12 crc kubenswrapper[4796]: I1127 11:54:12.675314 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:54:12 crc kubenswrapper[4796]: I1127 11:54:12.683085 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-zzhd9" Nov 27 11:54:12 crc kubenswrapper[4796]: I1127 11:54:12.905033 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:54:12 crc kubenswrapper[4796]: I1127 11:54:12.913039 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w" Nov 27 11:54:17 crc kubenswrapper[4796]: I1127 11:54:17.118399 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" Nov 27 11:54:17 crc kubenswrapper[4796]: I1127 11:54:17.451174 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6f77465788-gr9vn" Nov 27 11:54:18 crc kubenswrapper[4796]: I1127 11:54:18.569122 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:54:18 crc kubenswrapper[4796]: E1127 11:54:18.569425 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:54:30 crc kubenswrapper[4796]: I1127 11:54:30.568800 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:54:30 crc kubenswrapper[4796]: E1127 11:54:30.569478 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.715608 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r7r9l"] Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.717444 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.720798 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.721019 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.721174 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-csq67" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.721348 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.736402 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r7r9l"] Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.769073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31353d3-7382-407a-9e00-66117f32a415-config\") pod \"dnsmasq-dns-675f4bcbfc-r7r9l\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.769122 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n94s\" (UniqueName: \"kubernetes.io/projected/c31353d3-7382-407a-9e00-66117f32a415-kube-api-access-7n94s\") pod \"dnsmasq-dns-675f4bcbfc-r7r9l\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.787736 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s75hz"] Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.789110 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.795368 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.809209 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s75hz"] Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.869873 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31353d3-7382-407a-9e00-66117f32a415-config\") pod \"dnsmasq-dns-675f4bcbfc-r7r9l\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.870141 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n94s\" (UniqueName: \"kubernetes.io/projected/c31353d3-7382-407a-9e00-66117f32a415-kube-api-access-7n94s\") pod \"dnsmasq-dns-675f4bcbfc-r7r9l\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.870230 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.870345 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49r7h\" (UniqueName: \"kubernetes.io/projected/c308cf4b-8be9-4c7b-b059-51371552696b-kube-api-access-49r7h\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.870426 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-config\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.870781 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31353d3-7382-407a-9e00-66117f32a415-config\") pod \"dnsmasq-dns-675f4bcbfc-r7r9l\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.891242 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n94s\" (UniqueName: \"kubernetes.io/projected/c31353d3-7382-407a-9e00-66117f32a415-kube-api-access-7n94s\") pod \"dnsmasq-dns-675f4bcbfc-r7r9l\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.971173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.971500 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49r7h\" (UniqueName: \"kubernetes.io/projected/c308cf4b-8be9-4c7b-b059-51371552696b-kube-api-access-49r7h\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.971579 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-config\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.972725 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.973118 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-config\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:32 crc kubenswrapper[4796]: I1127 11:54:32.990063 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49r7h\" (UniqueName: \"kubernetes.io/projected/c308cf4b-8be9-4c7b-b059-51371552696b-kube-api-access-49r7h\") pod \"dnsmasq-dns-78dd6ddcc-s75hz\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:33 crc kubenswrapper[4796]: I1127 11:54:33.046374 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:54:33 crc kubenswrapper[4796]: I1127 11:54:33.121605 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:54:33 crc kubenswrapper[4796]: I1127 11:54:33.507093 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r7r9l"] Nov 27 11:54:33 crc kubenswrapper[4796]: I1127 11:54:33.620677 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s75hz"] Nov 27 11:54:33 crc kubenswrapper[4796]: W1127 11:54:33.623996 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc308cf4b_8be9_4c7b_b059_51371552696b.slice/crio-0ced6788e08b104140c280c9959c60189e9a67af28c1b34230a94eef1188fec3 WatchSource:0}: Error finding container 0ced6788e08b104140c280c9959c60189e9a67af28c1b34230a94eef1188fec3: Status 404 returned error can't find the container with id 0ced6788e08b104140c280c9959c60189e9a67af28c1b34230a94eef1188fec3 Nov 27 11:54:33 crc kubenswrapper[4796]: I1127 11:54:33.899632 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" event={"ID":"c31353d3-7382-407a-9e00-66117f32a415","Type":"ContainerStarted","Data":"415ce42f50d152da32f0fb447d36c02b92ca11a4e6f30af8e122176b00707156"} Nov 27 11:54:33 crc kubenswrapper[4796]: I1127 11:54:33.902498 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" event={"ID":"c308cf4b-8be9-4c7b-b059-51371552696b","Type":"ContainerStarted","Data":"0ced6788e08b104140c280c9959c60189e9a67af28c1b34230a94eef1188fec3"} Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.669410 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r7r9l"] Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.703160 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zxqll"] Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.704372 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.714159 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zxqll"] Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.827964 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.828008 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-config\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.828117 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm2c4\" (UniqueName: \"kubernetes.io/projected/8dec9dde-934b-49b9-ab26-3ce242e506f2-kube-api-access-pm2c4\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.929447 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.929502 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-config\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.929573 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm2c4\" (UniqueName: \"kubernetes.io/projected/8dec9dde-934b-49b9-ab26-3ce242e506f2-kube-api-access-pm2c4\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.930645 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.931164 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-config\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.945717 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s75hz"] Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.965601 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-n86r7"] Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.967629 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.970850 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm2c4\" (UniqueName: \"kubernetes.io/projected/8dec9dde-934b-49b9-ab26-3ce242e506f2-kube-api-access-pm2c4\") pod \"dnsmasq-dns-666b6646f7-zxqll\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:35 crc kubenswrapper[4796]: I1127 11:54:35.981780 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-n86r7"] Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.038715 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.134924 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.134974 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7ngd\" (UniqueName: \"kubernetes.io/projected/e56393a1-1d53-4e46-834a-c6cdd49bc31b-kube-api-access-f7ngd\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.135009 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-config\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.255066 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.255420 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7ngd\" (UniqueName: \"kubernetes.io/projected/e56393a1-1d53-4e46-834a-c6cdd49bc31b-kube-api-access-f7ngd\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.255453 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-config\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.256693 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-config\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.257042 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.281668 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7ngd\" (UniqueName: \"kubernetes.io/projected/e56393a1-1d53-4e46-834a-c6cdd49bc31b-kube-api-access-f7ngd\") pod \"dnsmasq-dns-57d769cc4f-n86r7\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.312520 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.680989 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zxqll"] Nov 27 11:54:36 crc kubenswrapper[4796]: W1127 11:54:36.684791 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dec9dde_934b_49b9_ab26_3ce242e506f2.slice/crio-57ae8b14ecb2ecea13da9313f1fd99cffe9d5cf13798feff45fc46b3d847c7c8 WatchSource:0}: Error finding container 57ae8b14ecb2ecea13da9313f1fd99cffe9d5cf13798feff45fc46b3d847c7c8: Status 404 returned error can't find the container with id 57ae8b14ecb2ecea13da9313f1fd99cffe9d5cf13798feff45fc46b3d847c7c8 Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.857602 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.860032 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.863010 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.863107 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.863369 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.864641 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.864864 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.865165 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h228j" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.865689 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.869676 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.869755 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.869791 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.871028 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.872004 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.872176 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.872312 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.872499 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.873411 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.874617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.874768 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq8jg\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-kube-api-access-nq8jg\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.874966 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.889185 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-n86r7"] Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.925660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" event={"ID":"8dec9dde-934b-49b9-ab26-3ce242e506f2","Type":"ContainerStarted","Data":"57ae8b14ecb2ecea13da9313f1fd99cffe9d5cf13798feff45fc46b3d847c7c8"} Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980535 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq8jg\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-kube-api-access-nq8jg\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980582 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980627 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980679 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980860 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980901 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.980935 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.981049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.981075 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.981117 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.981157 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.981324 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.982038 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.982943 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.983290 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.983403 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.984071 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.988404 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.988534 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.988844 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:36 crc kubenswrapper[4796]: I1127 11:54:36.996827 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.001713 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq8jg\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-kube-api-access-nq8jg\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.025421 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " pod="openstack/rabbitmq-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.107549 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.114746 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.116764 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.117640 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.120571 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.121567 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.121883 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.122021 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-b955p" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.122148 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.122153 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.188320 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285313 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285365 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e9cd74f-d96d-4be4-b448-01d027b34689-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285391 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjhvl\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-kube-api-access-cjhvl\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285496 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285658 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285690 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285718 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e9cd74f-d96d-4be4-b448-01d027b34689-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285769 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285792 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.285808 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388108 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e9cd74f-d96d-4be4-b448-01d027b34689-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388142 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388162 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388183 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388201 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388227 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e9cd74f-d96d-4be4-b448-01d027b34689-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388249 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjhvl\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-kube-api-access-cjhvl\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388283 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388310 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388349 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.388936 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.389403 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.389961 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.390770 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.391047 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.391439 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.392820 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e9cd74f-d96d-4be4-b448-01d027b34689-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.394680 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.394738 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e9cd74f-d96d-4be4-b448-01d027b34689-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.395138 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.408141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjhvl\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-kube-api-access-cjhvl\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.416104 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:37 crc kubenswrapper[4796]: I1127 11:54:37.749511 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.612491 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.618175 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.626860 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.656703 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.656749 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.656753 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8ts2r" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.656842 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.662791 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812211 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812532 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8f3e487f-6783-499b-bd57-0a811f4ef5f8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812564 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812580 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bds7\" (UniqueName: \"kubernetes.io/projected/8f3e487f-6783-499b-bd57-0a811f4ef5f8-kube-api-access-8bds7\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812609 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-kolla-config\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812650 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3e487f-6783-499b-bd57-0a811f4ef5f8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812675 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-config-data-default\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.812705 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f3e487f-6783-499b-bd57-0a811f4ef5f8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f3e487f-6783-499b-bd57-0a811f4ef5f8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914191 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914226 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8f3e487f-6783-499b-bd57-0a811f4ef5f8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914250 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914279 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bds7\" (UniqueName: \"kubernetes.io/projected/8f3e487f-6783-499b-bd57-0a811f4ef5f8-kube-api-access-8bds7\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914306 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-kolla-config\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914331 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3e487f-6783-499b-bd57-0a811f4ef5f8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.914354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-config-data-default\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.915145 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-config-data-default\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.915793 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.916377 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-kolla-config\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.917449 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8f3e487f-6783-499b-bd57-0a811f4ef5f8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.917955 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f3e487f-6783-499b-bd57-0a811f4ef5f8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.922873 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f3e487f-6783-499b-bd57-0a811f4ef5f8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.927155 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3e487f-6783-499b-bd57-0a811f4ef5f8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.944829 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.947516 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bds7\" (UniqueName: \"kubernetes.io/projected/8f3e487f-6783-499b-bd57-0a811f4ef5f8-kube-api-access-8bds7\") pod \"openstack-galera-0\" (UID: \"8f3e487f-6783-499b-bd57-0a811f4ef5f8\") " pod="openstack/openstack-galera-0" Nov 27 11:54:38 crc kubenswrapper[4796]: I1127 11:54:38.973248 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 27 11:54:39 crc kubenswrapper[4796]: I1127 11:54:39.984905 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 11:54:39 crc kubenswrapper[4796]: I1127 11:54:39.986836 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:39 crc kubenswrapper[4796]: I1127 11:54:39.993741 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-72k9j" Nov 27 11:54:39 crc kubenswrapper[4796]: I1127 11:54:39.993875 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 27 11:54:39 crc kubenswrapper[4796]: I1127 11:54:39.994616 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 27 11:54:39 crc kubenswrapper[4796]: I1127 11:54:39.994665 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.008290 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.134999 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135048 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135132 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135378 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135403 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135434 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135658 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.135739 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6vrc\" (UniqueName: \"kubernetes.io/projected/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-kube-api-access-h6vrc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.153620 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.154612 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.156004 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.156008 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.164781 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.168390 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-cf7vz" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237549 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6vrc\" (UniqueName: \"kubernetes.io/projected/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-kube-api-access-h6vrc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237629 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237655 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237672 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237724 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237740 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.237761 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.238061 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.238436 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.238548 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.238939 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.239607 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.242710 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.243093 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.256415 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6vrc\" (UniqueName: \"kubernetes.io/projected/bcc2f71a-d686-4c12-adc3-6a18ca6225f8-kube-api-access-h6vrc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.262793 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"bcc2f71a-d686-4c12-adc3-6a18ca6225f8\") " pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.317729 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.339878 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.339996 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.340087 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-config-data\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.340352 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-kolla-config\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.340423 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7xtk\" (UniqueName: \"kubernetes.io/projected/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-kube-api-access-m7xtk\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: W1127 11:54:40.357147 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode56393a1_1d53_4e46_834a_c6cdd49bc31b.slice/crio-24df41992c38e3993a46f0e3e2897b7d2fbb49286c1af7ec9849ff2971491453 WatchSource:0}: Error finding container 24df41992c38e3993a46f0e3e2897b7d2fbb49286c1af7ec9849ff2971491453: Status 404 returned error can't find the container with id 24df41992c38e3993a46f0e3e2897b7d2fbb49286c1af7ec9849ff2971491453 Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.441668 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-kolla-config\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.441730 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7xtk\" (UniqueName: \"kubernetes.io/projected/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-kube-api-access-m7xtk\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.441766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.441812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.441849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-config-data\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.443496 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-config-data\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.443734 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-kolla-config\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.446575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.446592 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.460785 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7xtk\" (UniqueName: \"kubernetes.io/projected/7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21-kube-api-access-m7xtk\") pod \"memcached-0\" (UID: \"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21\") " pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.476250 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 27 11:54:40 crc kubenswrapper[4796]: I1127 11:54:40.958113 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" event={"ID":"e56393a1-1d53-4e46-834a-c6cdd49bc31b","Type":"ContainerStarted","Data":"24df41992c38e3993a46f0e3e2897b7d2fbb49286c1af7ec9849ff2971491453"} Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.165184 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.166494 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.170580 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-pbfwr" Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.180785 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.267603 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twdrc\" (UniqueName: \"kubernetes.io/projected/c3f46f0d-069e-4527-a8a5-ca60f6022f9c-kube-api-access-twdrc\") pod \"kube-state-metrics-0\" (UID: \"c3f46f0d-069e-4527-a8a5-ca60f6022f9c\") " pod="openstack/kube-state-metrics-0" Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.368465 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twdrc\" (UniqueName: \"kubernetes.io/projected/c3f46f0d-069e-4527-a8a5-ca60f6022f9c-kube-api-access-twdrc\") pod \"kube-state-metrics-0\" (UID: \"c3f46f0d-069e-4527-a8a5-ca60f6022f9c\") " pod="openstack/kube-state-metrics-0" Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.387528 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twdrc\" (UniqueName: \"kubernetes.io/projected/c3f46f0d-069e-4527-a8a5-ca60f6022f9c-kube-api-access-twdrc\") pod \"kube-state-metrics-0\" (UID: \"c3f46f0d-069e-4527-a8a5-ca60f6022f9c\") " pod="openstack/kube-state-metrics-0" Nov 27 11:54:42 crc kubenswrapper[4796]: I1127 11:54:42.526715 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 11:54:43 crc kubenswrapper[4796]: I1127 11:54:43.580297 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:54:43 crc kubenswrapper[4796]: E1127 11:54:43.580820 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:54:43 crc kubenswrapper[4796]: I1127 11:54:43.611576 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.762139 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zjmwj"] Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.764431 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.768703 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6kbwt" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.769050 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.769231 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.773200 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7cdqc"] Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.775182 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.797349 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zjmwj"] Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.797391 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7cdqc"] Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941038 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-ovn-controller-tls-certs\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941116 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-run\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941160 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-lib\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941321 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-run\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941402 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-combined-ca-bundle\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941431 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-run-ovn\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941450 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-log-ovn\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941486 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdjbx\" (UniqueName: \"kubernetes.io/projected/a6cbbb73-5031-4929-b184-81a8cdbc0610-kube-api-access-pdjbx\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941802 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrnvm\" (UniqueName: \"kubernetes.io/projected/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-kube-api-access-lrnvm\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941931 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-etc-ovs\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941967 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-log\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.941998 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cbbb73-5031-4929-b184-81a8cdbc0610-scripts\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:45 crc kubenswrapper[4796]: I1127 11:54:45.942119 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-scripts\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.043218 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-run\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.043607 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-combined-ca-bundle\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.043712 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-run-ovn\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.043794 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-log-ovn\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.043875 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-run\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.043910 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-run-ovn\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044046 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdjbx\" (UniqueName: \"kubernetes.io/projected/a6cbbb73-5031-4929-b184-81a8cdbc0610-kube-api-access-pdjbx\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044141 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-var-log-ovn\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044302 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrnvm\" (UniqueName: \"kubernetes.io/projected/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-kube-api-access-lrnvm\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044415 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-etc-ovs\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044779 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cbbb73-5031-4929-b184-81a8cdbc0610-scripts\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044886 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-log\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.045146 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-scripts\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.045031 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-log\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.044825 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-etc-ovs\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.045734 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-ovn-controller-tls-certs\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.046216 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-run\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.046416 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-lib\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.046976 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-lib\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.047951 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-scripts\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.047994 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6cbbb73-5031-4929-b184-81a8cdbc0610-var-run\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.050524 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-combined-ca-bundle\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.053998 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-ovn-controller-tls-certs\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.061192 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6cbbb73-5031-4929-b184-81a8cdbc0610-scripts\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.064424 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdjbx\" (UniqueName: \"kubernetes.io/projected/a6cbbb73-5031-4929-b184-81a8cdbc0610-kube-api-access-pdjbx\") pod \"ovn-controller-ovs-7cdqc\" (UID: \"a6cbbb73-5031-4929-b184-81a8cdbc0610\") " pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.066414 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrnvm\" (UniqueName: \"kubernetes.io/projected/99ef6c07-3c37-4a97-8fbc-bd09a91617a0-kube-api-access-lrnvm\") pod \"ovn-controller-zjmwj\" (UID: \"99ef6c07-3c37-4a97-8fbc-bd09a91617a0\") " pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.129960 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.138233 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.691548 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.693240 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.698160 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.698370 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.698545 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.698610 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-52rrz" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.698791 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.710947 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861057 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861158 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861213 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/19c0f33c-82a7-4987-be98-e699c9d32051-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861282 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861317 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19c0f33c-82a7-4987-be98-e699c9d32051-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861347 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c0f33c-82a7-4987-be98-e699c9d32051-config\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861471 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.861533 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fztcv\" (UniqueName: \"kubernetes.io/projected/19c0f33c-82a7-4987-be98-e699c9d32051-kube-api-access-fztcv\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963085 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c0f33c-82a7-4987-be98-e699c9d32051-config\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963145 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fztcv\" (UniqueName: \"kubernetes.io/projected/19c0f33c-82a7-4987-be98-e699c9d32051-kube-api-access-fztcv\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963169 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963227 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963253 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963288 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/19c0f33c-82a7-4987-be98-e699c9d32051-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963326 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963357 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19c0f33c-82a7-4987-be98-e699c9d32051-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.963657 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.964013 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19c0f33c-82a7-4987-be98-e699c9d32051-config\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.964565 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19c0f33c-82a7-4987-be98-e699c9d32051-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.964637 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/19c0f33c-82a7-4987-be98-e699c9d32051-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.971472 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.972885 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.984653 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fztcv\" (UniqueName: \"kubernetes.io/projected/19c0f33c-82a7-4987-be98-e699c9d32051-kube-api-access-fztcv\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.989343 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c0f33c-82a7-4987-be98-e699c9d32051-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:46 crc kubenswrapper[4796]: I1127 11:54:46.992239 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"19c0f33c-82a7-4987-be98-e699c9d32051\") " pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:47 crc kubenswrapper[4796]: I1127 11:54:47.024690 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.910818 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.912477 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.916000 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.916101 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.916127 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k2jzn" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.916167 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.940643 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993784 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzrx\" (UniqueName: \"kubernetes.io/projected/8999de5c-3782-44dc-b88d-29f80c31a55f-kube-api-access-8kzrx\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993845 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993864 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999de5c-3782-44dc-b88d-29f80c31a55f-config\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993900 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8999de5c-3782-44dc-b88d-29f80c31a55f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993919 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993940 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993956 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999de5c-3782-44dc-b88d-29f80c31a55f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:48 crc kubenswrapper[4796]: I1127 11:54:48.993975 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095104 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzrx\" (UniqueName: \"kubernetes.io/projected/8999de5c-3782-44dc-b88d-29f80c31a55f-kube-api-access-8kzrx\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095174 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095202 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999de5c-3782-44dc-b88d-29f80c31a55f-config\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095248 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8999de5c-3782-44dc-b88d-29f80c31a55f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095307 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095338 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095361 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999de5c-3782-44dc-b88d-29f80c31a55f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.095392 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.096365 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8999de5c-3782-44dc-b88d-29f80c31a55f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.097975 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8999de5c-3782-44dc-b88d-29f80c31a55f-config\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.099706 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999de5c-3782-44dc-b88d-29f80c31a55f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.102115 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.106436 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.107516 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.108767 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8999de5c-3782-44dc-b88d-29f80c31a55f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.118217 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzrx\" (UniqueName: \"kubernetes.io/projected/8999de5c-3782-44dc-b88d-29f80c31a55f-kube-api-access-8kzrx\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.147171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8999de5c-3782-44dc-b88d-29f80c31a55f\") " pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:49 crc kubenswrapper[4796]: I1127 11:54:49.250330 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 27 11:54:55 crc kubenswrapper[4796]: I1127 11:54:55.569832 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:54:55 crc kubenswrapper[4796]: E1127 11:54:55.570624 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:54:59 crc kubenswrapper[4796]: I1127 11:54:59.463883 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 27 11:54:59 crc kubenswrapper[4796]: W1127 11:54:59.827283 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb853d8bd_9978_45a7_ad7c_b3217d4a93b3.slice/crio-a9001c375a0b55a1ac3a65d849e89280d854a1d87ca9594984d173bfd1a52585 WatchSource:0}: Error finding container a9001c375a0b55a1ac3a65d849e89280d854a1d87ca9594984d173bfd1a52585: Status 404 returned error can't find the container with id a9001c375a0b55a1ac3a65d849e89280d854a1d87ca9594984d173bfd1a52585 Nov 27 11:54:59 crc kubenswrapper[4796]: W1127 11:54:59.858127 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f3e487f_6783_499b_bd57_0a811f4ef5f8.slice/crio-6a499ca6dcd3a5bdbf9b0cb50e5f390e2525ee1e6d31d4549d0c3a108127b13b WatchSource:0}: Error finding container 6a499ca6dcd3a5bdbf9b0cb50e5f390e2525ee1e6d31d4549d0c3a108127b13b: Status 404 returned error can't find the container with id 6a499ca6dcd3a5bdbf9b0cb50e5f390e2525ee1e6d31d4549d0c3a108127b13b Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.872633 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.872863 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pm2c4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-zxqll_openstack(8dec9dde-934b-49b9-ab26-3ce242e506f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.874103 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" podUID="8dec9dde-934b-49b9-ab26-3ce242e506f2" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.905656 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.905817 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7n94s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-r7r9l_openstack(c31353d3-7382-407a-9e00-66117f32a415): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.907406 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.907503 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-49r7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-s75hz_openstack(c308cf4b-8be9-4c7b-b059-51371552696b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.907578 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" podUID="c31353d3-7382-407a-9e00-66117f32a415" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.908642 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" podUID="c308cf4b-8be9-4c7b-b059-51371552696b" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.953701 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.954023 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f7ngd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-n86r7_openstack(e56393a1-1d53-4e46-834a-c6cdd49bc31b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:54:59 crc kubenswrapper[4796]: E1127 11:54:59.955299 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" podUID="e56393a1-1d53-4e46-834a-c6cdd49bc31b" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.119424 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b853d8bd-9978-45a7-ad7c-b3217d4a93b3","Type":"ContainerStarted","Data":"a9001c375a0b55a1ac3a65d849e89280d854a1d87ca9594984d173bfd1a52585"} Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.128538 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8f3e487f-6783-499b-bd57-0a811f4ef5f8","Type":"ContainerStarted","Data":"6a499ca6dcd3a5bdbf9b0cb50e5f390e2525ee1e6d31d4549d0c3a108127b13b"} Nov 27 11:55:00 crc kubenswrapper[4796]: E1127 11:55:00.132177 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" podUID="8dec9dde-934b-49b9-ab26-3ce242e506f2" Nov 27 11:55:00 crc kubenswrapper[4796]: E1127 11:55:00.132409 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" podUID="e56393a1-1d53-4e46-834a-c6cdd49bc31b" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.697889 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.699249 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.713783 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.737889 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.748486 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.747140 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.812942 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-config\") pod \"c308cf4b-8be9-4c7b-b059-51371552696b\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.813076 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n94s\" (UniqueName: \"kubernetes.io/projected/c31353d3-7382-407a-9e00-66117f32a415-kube-api-access-7n94s\") pod \"c31353d3-7382-407a-9e00-66117f32a415\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.813122 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49r7h\" (UniqueName: \"kubernetes.io/projected/c308cf4b-8be9-4c7b-b059-51371552696b-kube-api-access-49r7h\") pod \"c308cf4b-8be9-4c7b-b059-51371552696b\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.813205 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31353d3-7382-407a-9e00-66117f32a415-config\") pod \"c31353d3-7382-407a-9e00-66117f32a415\" (UID: \"c31353d3-7382-407a-9e00-66117f32a415\") " Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.813419 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-dns-svc\") pod \"c308cf4b-8be9-4c7b-b059-51371552696b\" (UID: \"c308cf4b-8be9-4c7b-b059-51371552696b\") " Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.814372 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-config" (OuterVolumeSpecName: "config") pod "c308cf4b-8be9-4c7b-b059-51371552696b" (UID: "c308cf4b-8be9-4c7b-b059-51371552696b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.814386 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c308cf4b-8be9-4c7b-b059-51371552696b" (UID: "c308cf4b-8be9-4c7b-b059-51371552696b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.815114 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31353d3-7382-407a-9e00-66117f32a415-config" (OuterVolumeSpecName: "config") pod "c31353d3-7382-407a-9e00-66117f32a415" (UID: "c31353d3-7382-407a-9e00-66117f32a415"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.821780 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31353d3-7382-407a-9e00-66117f32a415-kube-api-access-7n94s" (OuterVolumeSpecName: "kube-api-access-7n94s") pod "c31353d3-7382-407a-9e00-66117f32a415" (UID: "c31353d3-7382-407a-9e00-66117f32a415"). InnerVolumeSpecName "kube-api-access-7n94s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.821862 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c308cf4b-8be9-4c7b-b059-51371552696b-kube-api-access-49r7h" (OuterVolumeSpecName: "kube-api-access-49r7h") pod "c308cf4b-8be9-4c7b-b059-51371552696b" (UID: "c308cf4b-8be9-4c7b-b059-51371552696b"). InnerVolumeSpecName "kube-api-access-49r7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.851208 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7cdqc"] Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.915176 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.915213 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308cf4b-8be9-4c7b-b059-51371552696b-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.915223 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n94s\" (UniqueName: \"kubernetes.io/projected/c31353d3-7382-407a-9e00-66117f32a415-kube-api-access-7n94s\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.915235 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49r7h\" (UniqueName: \"kubernetes.io/projected/c308cf4b-8be9-4c7b-b059-51371552696b-kube-api-access-49r7h\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.915243 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31353d3-7382-407a-9e00-66117f32a415-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:00 crc kubenswrapper[4796]: I1127 11:55:00.977701 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zjmwj"] Nov 27 11:55:00 crc kubenswrapper[4796]: W1127 11:55:00.992518 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99ef6c07_3c37_4a97_8fbc_bd09a91617a0.slice/crio-273f83bb4a92ab2781c055a04317f3d7a7730b403feefb5aa85428742d34dca1 WatchSource:0}: Error finding container 273f83bb4a92ab2781c055a04317f3d7a7730b403feefb5aa85428742d34dca1: Status 404 returned error can't find the container with id 273f83bb4a92ab2781c055a04317f3d7a7730b403feefb5aa85428742d34dca1 Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.064639 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.137298 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3f46f0d-069e-4527-a8a5-ca60f6022f9c","Type":"ContainerStarted","Data":"9e67cecbcf04d09da4fe90137c7042fd3a1d52ce4ad34281ff09d1a43871738f"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.138481 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e9cd74f-d96d-4be4-b448-01d027b34689","Type":"ContainerStarted","Data":"2d1a4349c8fa3954bd9222adf0dd39a1ffea03fc072dd421a2afaaba355dc2f6"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.139551 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" event={"ID":"c308cf4b-8be9-4c7b-b059-51371552696b","Type":"ContainerDied","Data":"0ced6788e08b104140c280c9959c60189e9a67af28c1b34230a94eef1188fec3"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.139574 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-s75hz" Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.141740 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj" event={"ID":"99ef6c07-3c37-4a97-8fbc-bd09a91617a0","Type":"ContainerStarted","Data":"273f83bb4a92ab2781c055a04317f3d7a7730b403feefb5aa85428742d34dca1"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.143048 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"19c0f33c-82a7-4987-be98-e699c9d32051","Type":"ContainerStarted","Data":"b7fd242ce8e36455470eab1c74e401e6ddf89627ac12816fdadacf6e3832333e"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.144250 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.144251 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-r7r9l" event={"ID":"c31353d3-7382-407a-9e00-66117f32a415","Type":"ContainerDied","Data":"415ce42f50d152da32f0fb447d36c02b92ca11a4e6f30af8e122176b00707156"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.145495 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7cdqc" event={"ID":"a6cbbb73-5031-4929-b184-81a8cdbc0610","Type":"ContainerStarted","Data":"61301e2cdb12605b7e86de2405abc7efcdd799b62d7c3e720dfa7c342532bb9d"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.147083 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bcc2f71a-d686-4c12-adc3-6a18ca6225f8","Type":"ContainerStarted","Data":"497809ca38ee1d9399262fba92b4bbcaec5110c175b7367c415a290a79ba85f7"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.148379 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21","Type":"ContainerStarted","Data":"ae8f6d5c43f0b9191097903fedacbd7d38e64bec834e4252cbda93cf381673b0"} Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.256163 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s75hz"] Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.262192 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-s75hz"] Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.274357 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r7r9l"] Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.281046 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-r7r9l"] Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.578180 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c308cf4b-8be9-4c7b-b059-51371552696b" path="/var/lib/kubelet/pods/c308cf4b-8be9-4c7b-b059-51371552696b/volumes" Nov 27 11:55:01 crc kubenswrapper[4796]: I1127 11:55:01.578607 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31353d3-7382-407a-9e00-66117f32a415" path="/var/lib/kubelet/pods/c31353d3-7382-407a-9e00-66117f32a415/volumes" Nov 27 11:55:02 crc kubenswrapper[4796]: I1127 11:55:02.051051 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 11:55:02 crc kubenswrapper[4796]: W1127 11:55:02.116762 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8999de5c_3782_44dc_b88d_29f80c31a55f.slice/crio-7cd166546923a0fe7f450d521b9ffdfa379c34d50b7c19f46ed0399737ed4616 WatchSource:0}: Error finding container 7cd166546923a0fe7f450d521b9ffdfa379c34d50b7c19f46ed0399737ed4616: Status 404 returned error can't find the container with id 7cd166546923a0fe7f450d521b9ffdfa379c34d50b7c19f46ed0399737ed4616 Nov 27 11:55:02 crc kubenswrapper[4796]: I1127 11:55:02.162660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8999de5c-3782-44dc-b88d-29f80c31a55f","Type":"ContainerStarted","Data":"7cd166546923a0fe7f450d521b9ffdfa379c34d50b7c19f46ed0399737ed4616"} Nov 27 11:55:08 crc kubenswrapper[4796]: I1127 11:55:08.568769 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:55:08 crc kubenswrapper[4796]: E1127 11:55:08.569672 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.196706 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-q2vvb"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.197748 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.200775 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.224227 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-q2vvb"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.300168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5cd65c9-3282-4039-9963-f1174a310831-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.300246 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cd65c9-3282-4039-9963-f1174a310831-combined-ca-bundle\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.300321 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d5cd65c9-3282-4039-9963-f1174a310831-ovn-rundir\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.300349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5cd65c9-3282-4039-9963-f1174a310831-config\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.300502 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d5cd65c9-3282-4039-9963-f1174a310831-ovs-rundir\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.300590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr6jc\" (UniqueName: \"kubernetes.io/projected/d5cd65c9-3282-4039-9963-f1174a310831-kube-api-access-cr6jc\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.332058 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-n86r7"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.357935 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9c62v"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.359687 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.364699 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.382988 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9c62v"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.402041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr6jc\" (UniqueName: \"kubernetes.io/projected/d5cd65c9-3282-4039-9963-f1174a310831-kube-api-access-cr6jc\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.402120 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5cd65c9-3282-4039-9963-f1174a310831-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.402170 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cd65c9-3282-4039-9963-f1174a310831-combined-ca-bundle\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.402195 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d5cd65c9-3282-4039-9963-f1174a310831-ovn-rundir\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.402221 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5cd65c9-3282-4039-9963-f1174a310831-config\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.402251 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d5cd65c9-3282-4039-9963-f1174a310831-ovs-rundir\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.403317 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d5cd65c9-3282-4039-9963-f1174a310831-ovs-rundir\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.403528 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d5cd65c9-3282-4039-9963-f1174a310831-ovn-rundir\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.404371 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5cd65c9-3282-4039-9963-f1174a310831-config\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.411781 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5cd65c9-3282-4039-9963-f1174a310831-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.415406 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cd65c9-3282-4039-9963-f1174a310831-combined-ca-bundle\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.439344 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr6jc\" (UniqueName: \"kubernetes.io/projected/d5cd65c9-3282-4039-9963-f1174a310831-kube-api-access-cr6jc\") pod \"ovn-controller-metrics-q2vvb\" (UID: \"d5cd65c9-3282-4039-9963-f1174a310831\") " pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.503775 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.503965 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-config\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.504310 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xts4b\" (UniqueName: \"kubernetes.io/projected/00d9b05c-b63f-451b-bda5-1451a05ee320-kube-api-access-xts4b\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.504391 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.518069 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-q2vvb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.549286 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zxqll"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.596050 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnwth"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.597249 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.602339 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.605474 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-config\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.605535 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xts4b\" (UniqueName: \"kubernetes.io/projected/00d9b05c-b63f-451b-bda5-1451a05ee320-kube-api-access-xts4b\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.605557 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.605603 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.606399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.606986 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.606985 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-config\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.625727 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xts4b\" (UniqueName: \"kubernetes.io/projected/00d9b05c-b63f-451b-bda5-1451a05ee320-kube-api-access-xts4b\") pod \"dnsmasq-dns-6bc7876d45-9c62v\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.628844 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnwth"] Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.684867 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.707097 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b7x4\" (UniqueName: \"kubernetes.io/projected/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-kube-api-access-2b7x4\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.707172 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-config\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.707250 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-dns-svc\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.707325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.707408 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.808966 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.809077 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b7x4\" (UniqueName: \"kubernetes.io/projected/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-kube-api-access-2b7x4\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.809106 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-config\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.809134 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-dns-svc\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.809165 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.810231 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.810258 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.810389 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-config\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.810675 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-dns-svc\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.834461 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b7x4\" (UniqueName: \"kubernetes.io/projected/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-kube-api-access-2b7x4\") pod \"dnsmasq-dns-8554648995-mnwth\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:09 crc kubenswrapper[4796]: I1127 11:55:09.917483 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.609029 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.615805 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.743819 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-dns-svc\") pod \"8dec9dde-934b-49b9-ab26-3ce242e506f2\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.743871 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm2c4\" (UniqueName: \"kubernetes.io/projected/8dec9dde-934b-49b9-ab26-3ce242e506f2-kube-api-access-pm2c4\") pod \"8dec9dde-934b-49b9-ab26-3ce242e506f2\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.744005 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-config\") pod \"8dec9dde-934b-49b9-ab26-3ce242e506f2\" (UID: \"8dec9dde-934b-49b9-ab26-3ce242e506f2\") " Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.744051 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-config\") pod \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.744110 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-dns-svc\") pod \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.744246 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7ngd\" (UniqueName: \"kubernetes.io/projected/e56393a1-1d53-4e46-834a-c6cdd49bc31b-kube-api-access-f7ngd\") pod \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\" (UID: \"e56393a1-1d53-4e46-834a-c6cdd49bc31b\") " Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.745126 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8dec9dde-934b-49b9-ab26-3ce242e506f2" (UID: "8dec9dde-934b-49b9-ab26-3ce242e506f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.747461 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-config" (OuterVolumeSpecName: "config") pod "e56393a1-1d53-4e46-834a-c6cdd49bc31b" (UID: "e56393a1-1d53-4e46-834a-c6cdd49bc31b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.747466 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-config" (OuterVolumeSpecName: "config") pod "8dec9dde-934b-49b9-ab26-3ce242e506f2" (UID: "8dec9dde-934b-49b9-ab26-3ce242e506f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.747486 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e56393a1-1d53-4e46-834a-c6cdd49bc31b" (UID: "e56393a1-1d53-4e46-834a-c6cdd49bc31b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.749257 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e56393a1-1d53-4e46-834a-c6cdd49bc31b-kube-api-access-f7ngd" (OuterVolumeSpecName: "kube-api-access-f7ngd") pod "e56393a1-1d53-4e46-834a-c6cdd49bc31b" (UID: "e56393a1-1d53-4e46-834a-c6cdd49bc31b"). InnerVolumeSpecName "kube-api-access-f7ngd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.751883 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dec9dde-934b-49b9-ab26-3ce242e506f2-kube-api-access-pm2c4" (OuterVolumeSpecName: "kube-api-access-pm2c4") pod "8dec9dde-934b-49b9-ab26-3ce242e506f2" (UID: "8dec9dde-934b-49b9-ab26-3ce242e506f2"). InnerVolumeSpecName "kube-api-access-pm2c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.846052 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.846082 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.846091 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56393a1-1d53-4e46-834a-c6cdd49bc31b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.846102 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7ngd\" (UniqueName: \"kubernetes.io/projected/e56393a1-1d53-4e46-834a-c6cdd49bc31b-kube-api-access-f7ngd\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.846114 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8dec9dde-934b-49b9-ab26-3ce242e506f2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:11 crc kubenswrapper[4796]: I1127 11:55:11.846124 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm2c4\" (UniqueName: \"kubernetes.io/projected/8dec9dde-934b-49b9-ab26-3ce242e506f2-kube-api-access-pm2c4\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.246816 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" event={"ID":"e56393a1-1d53-4e46-834a-c6cdd49bc31b","Type":"ContainerDied","Data":"24df41992c38e3993a46f0e3e2897b7d2fbb49286c1af7ec9849ff2971491453"} Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.246843 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-n86r7" Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.248470 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" event={"ID":"8dec9dde-934b-49b9-ab26-3ce242e506f2","Type":"ContainerDied","Data":"57ae8b14ecb2ecea13da9313f1fd99cffe9d5cf13798feff45fc46b3d847c7c8"} Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.248590 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zxqll" Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.319521 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-n86r7"] Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.326229 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-n86r7"] Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.342550 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zxqll"] Nov 27 11:55:12 crc kubenswrapper[4796]: I1127 11:55:12.348278 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zxqll"] Nov 27 11:55:13 crc kubenswrapper[4796]: I1127 11:55:13.590728 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dec9dde-934b-49b9-ab26-3ce242e506f2" path="/var/lib/kubelet/pods/8dec9dde-934b-49b9-ab26-3ce242e506f2/volumes" Nov 27 11:55:13 crc kubenswrapper[4796]: I1127 11:55:13.592759 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e56393a1-1d53-4e46-834a-c6cdd49bc31b" path="/var/lib/kubelet/pods/e56393a1-1d53-4e46-834a-c6cdd49bc31b/volumes" Nov 27 11:55:14 crc kubenswrapper[4796]: I1127 11:55:14.321920 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9c62v"] Nov 27 11:55:14 crc kubenswrapper[4796]: I1127 11:55:14.329062 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-q2vvb"] Nov 27 11:55:14 crc kubenswrapper[4796]: I1127 11:55:14.428588 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnwth"] Nov 27 11:55:15 crc kubenswrapper[4796]: I1127 11:55:15.295874 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-q2vvb" event={"ID":"d5cd65c9-3282-4039-9963-f1174a310831","Type":"ContainerStarted","Data":"6b07a9959cc8be93a1195e5d5919b74b4ab4beae7019e13db8cd6cc810f6ee62"} Nov 27 11:55:15 crc kubenswrapper[4796]: I1127 11:55:15.298532 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnwth" event={"ID":"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940","Type":"ContainerStarted","Data":"21994199e6d9e358a1c0153515b8a6ba5e7f2331027796e7f751e2963caad517"} Nov 27 11:55:15 crc kubenswrapper[4796]: I1127 11:55:15.300202 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" event={"ID":"00d9b05c-b63f-451b-bda5-1451a05ee320","Type":"ContainerStarted","Data":"a793f1bffcd11ef856df7a016163fc3d4d37f69a537ccb570822d75c1179ce66"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.312239 4796 generic.go:334] "Generic (PLEG): container finished" podID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerID="796ff09375965554e3face8ac11beca006bce3c45eec55671118b5d267a19796" exitCode=0 Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.312305 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" event={"ID":"00d9b05c-b63f-451b-bda5-1451a05ee320","Type":"ContainerDied","Data":"796ff09375965554e3face8ac11beca006bce3c45eec55671118b5d267a19796"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.328904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj" event={"ID":"99ef6c07-3c37-4a97-8fbc-bd09a91617a0","Type":"ContainerStarted","Data":"b5d4ee026c964e9424d130f55051862fffe4f2b05fae87d598aa31c239dd8cce"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.330104 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zjmwj" Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.337817 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"19c0f33c-82a7-4987-be98-e699c9d32051","Type":"ContainerStarted","Data":"3a24387f96281e1c4d74b93ff8323a59a1a4e996532ece2029518b9cff0566e5"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.343832 4796 generic.go:334] "Generic (PLEG): container finished" podID="a6cbbb73-5031-4929-b184-81a8cdbc0610" containerID="bc86269fd4d6667053d1b066e49be3dba9dbcb5eebc50999f0dc3ea4cd886879" exitCode=0 Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.343938 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7cdqc" event={"ID":"a6cbbb73-5031-4929-b184-81a8cdbc0610","Type":"ContainerDied","Data":"bc86269fd4d6667053d1b066e49be3dba9dbcb5eebc50999f0dc3ea4cd886879"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.346369 4796 generic.go:334] "Generic (PLEG): container finished" podID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerID="b26b517c154e56b5dd4c04cc5b2a15216138edf9e963ef2910f6bf75c2345426" exitCode=0 Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.346421 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnwth" event={"ID":"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940","Type":"ContainerDied","Data":"b26b517c154e56b5dd4c04cc5b2a15216138edf9e963ef2910f6bf75c2345426"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.355918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8f3e487f-6783-499b-bd57-0a811f4ef5f8","Type":"ContainerStarted","Data":"65eb5675345facb7ebd54d4923e5fb16c63c5357b79d666f812646993c61944d"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.361479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e9cd74f-d96d-4be4-b448-01d027b34689","Type":"ContainerStarted","Data":"5d683f5dd13d49637f10b186e2b1c13b5c81e507a7e715a4253a8484afc9e8c6"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.367679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8999de5c-3782-44dc-b88d-29f80c31a55f","Type":"ContainerStarted","Data":"95fc68b152719ba2d16e2eba0ef59712ae0191f459093299c87d00f9e45a5636"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.373536 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bcc2f71a-d686-4c12-adc3-6a18ca6225f8","Type":"ContainerStarted","Data":"3611f7076325ba443c692da034446ed79b676d9f033118c86939612b79d03d64"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.388251 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3f46f0d-069e-4527-a8a5-ca60f6022f9c","Type":"ContainerStarted","Data":"2ea395df8fdccd851176858cf9e794dcec4fef06c99a2aa68873cbdbf79f88e5"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.389385 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.390711 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21","Type":"ContainerStarted","Data":"17f03f0e91b8dac43a8b8d103239ad13767a35cc8ab3d417f5d0afb925504c13"} Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.390976 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.431553 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zjmwj" podStartSLOduration=18.969266482 podStartE2EDuration="31.431518047s" podCreationTimestamp="2025-11-27 11:54:45 +0000 UTC" firstStartedPulling="2025-11-27 11:55:00.995289846 +0000 UTC m=+1818.513608764" lastFinishedPulling="2025-11-27 11:55:13.457541371 +0000 UTC m=+1830.975860329" observedRunningTime="2025-11-27 11:55:16.355117449 +0000 UTC m=+1833.873436377" watchObservedRunningTime="2025-11-27 11:55:16.431518047 +0000 UTC m=+1833.949836975" Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.468683 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=20.131152488 podStartE2EDuration="34.468653488s" podCreationTimestamp="2025-11-27 11:54:42 +0000 UTC" firstStartedPulling="2025-11-27 11:55:00.74654328 +0000 UTC m=+1818.264862218" lastFinishedPulling="2025-11-27 11:55:15.0840443 +0000 UTC m=+1832.602363218" observedRunningTime="2025-11-27 11:55:16.459257117 +0000 UTC m=+1833.977576035" watchObservedRunningTime="2025-11-27 11:55:16.468653488 +0000 UTC m=+1833.986972406" Nov 27 11:55:16 crc kubenswrapper[4796]: I1127 11:55:16.476083 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=26.669099845 podStartE2EDuration="36.476066105s" podCreationTimestamp="2025-11-27 11:54:40 +0000 UTC" firstStartedPulling="2025-11-27 11:55:00.749434217 +0000 UTC m=+1818.267753135" lastFinishedPulling="2025-11-27 11:55:10.556400477 +0000 UTC m=+1828.074719395" observedRunningTime="2025-11-27 11:55:16.474568696 +0000 UTC m=+1833.992887614" watchObservedRunningTime="2025-11-27 11:55:16.476066105 +0000 UTC m=+1833.994385023" Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.401830 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" event={"ID":"00d9b05c-b63f-451b-bda5-1451a05ee320","Type":"ContainerStarted","Data":"e308b50574cab3eb6af00d23caf72872d8887a642827246b96ecfeec2233706e"} Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.402196 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.406218 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7cdqc" event={"ID":"a6cbbb73-5031-4929-b184-81a8cdbc0610","Type":"ContainerStarted","Data":"3966bc881f232d0c103642985fbee4d6e657753a1ea2bf6d6cc97601ec0d5613"} Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.406261 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7cdqc" event={"ID":"a6cbbb73-5031-4929-b184-81a8cdbc0610","Type":"ContainerStarted","Data":"1894cfe754d68b46aace097c758f5e250b71d84852b28482c50c807e9a578038"} Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.406296 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.406309 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.409734 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b853d8bd-9978-45a7-ad7c-b3217d4a93b3","Type":"ContainerStarted","Data":"c189a5bb7656fc2b2a2ccbdbd7bc017c1a6d0494b20cf5077e9aa8b7abe84c46"} Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.411742 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnwth" event={"ID":"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940","Type":"ContainerStarted","Data":"3ad28e41a17204998b025ee5336700e2f49ffd055f36804a31829a269e5f49dc"} Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.428077 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" podStartSLOduration=7.712301546 podStartE2EDuration="8.428057701s" podCreationTimestamp="2025-11-27 11:55:09 +0000 UTC" firstStartedPulling="2025-11-27 11:55:14.784477708 +0000 UTC m=+1832.302796626" lastFinishedPulling="2025-11-27 11:55:15.500233863 +0000 UTC m=+1833.018552781" observedRunningTime="2025-11-27 11:55:17.421748873 +0000 UTC m=+1834.940067791" watchObservedRunningTime="2025-11-27 11:55:17.428057701 +0000 UTC m=+1834.946376619" Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.443059 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7cdqc" podStartSLOduration=20.48586371 podStartE2EDuration="32.44303037s" podCreationTimestamp="2025-11-27 11:54:45 +0000 UTC" firstStartedPulling="2025-11-27 11:55:00.857145751 +0000 UTC m=+1818.375464669" lastFinishedPulling="2025-11-27 11:55:12.814312381 +0000 UTC m=+1830.332631329" observedRunningTime="2025-11-27 11:55:17.439006874 +0000 UTC m=+1834.957325802" watchObservedRunningTime="2025-11-27 11:55:17.44303037 +0000 UTC m=+1834.961349288" Nov 27 11:55:17 crc kubenswrapper[4796]: I1127 11:55:17.472361 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-mnwth" podStartSLOduration=7.757935595 podStartE2EDuration="8.472346443s" podCreationTimestamp="2025-11-27 11:55:09 +0000 UTC" firstStartedPulling="2025-11-27 11:55:14.784085448 +0000 UTC m=+1832.302404366" lastFinishedPulling="2025-11-27 11:55:15.498496296 +0000 UTC m=+1833.016815214" observedRunningTime="2025-11-27 11:55:17.464432432 +0000 UTC m=+1834.982751350" watchObservedRunningTime="2025-11-27 11:55:17.472346443 +0000 UTC m=+1834.990665351" Nov 27 11:55:18 crc kubenswrapper[4796]: I1127 11:55:18.422614 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.452956 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8999de5c-3782-44dc-b88d-29f80c31a55f","Type":"ContainerStarted","Data":"0ce6686b17d30a9fee621b3e417f5b884cc906c7614ff38033dd6f24c55a9bbf"} Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.456781 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"19c0f33c-82a7-4987-be98-e699c9d32051","Type":"ContainerStarted","Data":"951e8efda281b0e47092fcbe0364d9e826bc15d8fed434534e0b7e073ecd1bce"} Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.458494 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-q2vvb" event={"ID":"d5cd65c9-3282-4039-9963-f1174a310831","Type":"ContainerStarted","Data":"6ccc9f20e328510926f84780a066367da5050a04c407f7ae488af347f9a77418"} Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.460094 4796 generic.go:334] "Generic (PLEG): container finished" podID="bcc2f71a-d686-4c12-adc3-6a18ca6225f8" containerID="3611f7076325ba443c692da034446ed79b676d9f033118c86939612b79d03d64" exitCode=0 Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.460218 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bcc2f71a-d686-4c12-adc3-6a18ca6225f8","Type":"ContainerDied","Data":"3611f7076325ba443c692da034446ed79b676d9f033118c86939612b79d03d64"} Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.461995 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f3e487f-6783-499b-bd57-0a811f4ef5f8" containerID="65eb5675345facb7ebd54d4923e5fb16c63c5357b79d666f812646993c61944d" exitCode=0 Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.462030 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8f3e487f-6783-499b-bd57-0a811f4ef5f8","Type":"ContainerDied","Data":"65eb5675345facb7ebd54d4923e5fb16c63c5357b79d666f812646993c61944d"} Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.476242 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=16.526036968 podStartE2EDuration="33.476214337s" podCreationTimestamp="2025-11-27 11:54:47 +0000 UTC" firstStartedPulling="2025-11-27 11:55:02.119251129 +0000 UTC m=+1819.637570047" lastFinishedPulling="2025-11-27 11:55:19.069428498 +0000 UTC m=+1836.587747416" observedRunningTime="2025-11-27 11:55:20.470520845 +0000 UTC m=+1837.988839763" watchObservedRunningTime="2025-11-27 11:55:20.476214337 +0000 UTC m=+1837.994533295" Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.481459 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.552561 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=17.585691014 podStartE2EDuration="35.552543213s" podCreationTimestamp="2025-11-27 11:54:45 +0000 UTC" firstStartedPulling="2025-11-27 11:55:01.083745176 +0000 UTC m=+1818.602064094" lastFinishedPulling="2025-11-27 11:55:19.050597375 +0000 UTC m=+1836.568916293" observedRunningTime="2025-11-27 11:55:20.54906336 +0000 UTC m=+1838.067382288" watchObservedRunningTime="2025-11-27 11:55:20.552543213 +0000 UTC m=+1838.070862131" Nov 27 11:55:20 crc kubenswrapper[4796]: I1127 11:55:20.564978 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-q2vvb" podStartSLOduration=7.272445454 podStartE2EDuration="11.564958684s" podCreationTimestamp="2025-11-27 11:55:09 +0000 UTC" firstStartedPulling="2025-11-27 11:55:14.799915691 +0000 UTC m=+1832.318234629" lastFinishedPulling="2025-11-27 11:55:19.092428941 +0000 UTC m=+1836.610747859" observedRunningTime="2025-11-27 11:55:20.563376962 +0000 UTC m=+1838.081695880" watchObservedRunningTime="2025-11-27 11:55:20.564958684 +0000 UTC m=+1838.083277602" Nov 27 11:55:21 crc kubenswrapper[4796]: I1127 11:55:21.472955 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bcc2f71a-d686-4c12-adc3-6a18ca6225f8","Type":"ContainerStarted","Data":"6faf1093bc1dda5e9173a12ea03546b00ca547a3fe5cd34aef3a799262988a29"} Nov 27 11:55:21 crc kubenswrapper[4796]: I1127 11:55:21.475495 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8f3e487f-6783-499b-bd57-0a811f4ef5f8","Type":"ContainerStarted","Data":"7dabb87f4ebdc6f34686b6b6d1e4c9f7f1e8890e7c5402d5cd08b82a701f8199"} Nov 27 11:55:21 crc kubenswrapper[4796]: I1127 11:55:21.540053 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=30.697532447 podStartE2EDuration="43.540033046s" podCreationTimestamp="2025-11-27 11:54:38 +0000 UTC" firstStartedPulling="2025-11-27 11:55:00.746622402 +0000 UTC m=+1818.264941330" lastFinishedPulling="2025-11-27 11:55:13.589123011 +0000 UTC m=+1831.107441929" observedRunningTime="2025-11-27 11:55:21.507283732 +0000 UTC m=+1839.025602650" watchObservedRunningTime="2025-11-27 11:55:21.540033046 +0000 UTC m=+1839.058351964" Nov 27 11:55:21 crc kubenswrapper[4796]: I1127 11:55:21.542732 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=32.861607372 podStartE2EDuration="44.542719328s" podCreationTimestamp="2025-11-27 11:54:37 +0000 UTC" firstStartedPulling="2025-11-27 11:54:59.86262521 +0000 UTC m=+1817.380944148" lastFinishedPulling="2025-11-27 11:55:11.543737176 +0000 UTC m=+1829.062056104" observedRunningTime="2025-11-27 11:55:21.532733001 +0000 UTC m=+1839.051051949" watchObservedRunningTime="2025-11-27 11:55:21.542719328 +0000 UTC m=+1839.061038246" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.025359 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.251471 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.295472 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.482284 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.504861 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9c62v"] Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.505081 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="dnsmasq-dns" containerID="cri-o://e308b50574cab3eb6af00d23caf72872d8887a642827246b96ecfeec2233706e" gracePeriod=10 Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.509425 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.531745 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.532053 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.564338 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jpzwj"] Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.565684 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.570762 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:55:22 crc kubenswrapper[4796]: E1127 11:55:22.570997 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.588705 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jpzwj"] Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.700066 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.700151 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8cpp\" (UniqueName: \"kubernetes.io/projected/d7630686-df2d-484c-81d3-769308420b18-kube-api-access-c8cpp\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.700178 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.701093 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.701197 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-config\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.802505 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8cpp\" (UniqueName: \"kubernetes.io/projected/d7630686-df2d-484c-81d3-769308420b18-kube-api-access-c8cpp\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.802559 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.802589 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.802638 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-config\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.802696 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.803507 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.804250 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.804842 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.805474 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-config\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.825156 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8cpp\" (UniqueName: \"kubernetes.io/projected/d7630686-df2d-484c-81d3-769308420b18-kube-api-access-c8cpp\") pod \"dnsmasq-dns-b8fbc5445-jpzwj\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:22 crc kubenswrapper[4796]: I1127 11:55:22.950388 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.025341 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.089524 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.450069 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jpzwj"] Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.492164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" event={"ID":"d7630686-df2d-484c-81d3-769308420b18","Type":"ContainerStarted","Data":"1ba0998975266f85581e3b7f86c103289eae73b85f6b9d5f3dd53d1c92d8c970"} Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.530557 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.681686 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.700102 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.709798 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.709828 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lqbx8" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.710068 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.724626 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.758607 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.823189 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.823255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.823305 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/56f6dd56-8bbf-4610-b48c-898cda2e2aab-lock\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.823353 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdbbx\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-kube-api-access-wdbbx\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.823389 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/56f6dd56-8bbf-4610-b48c-898cda2e2aab-cache\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.829297 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.830872 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.833157 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.833433 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.833564 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-jvjh4" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.834168 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.847715 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925105 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdbbx\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-kube-api-access-wdbbx\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925433 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/56f6dd56-8bbf-4610-b48c-898cda2e2aab-cache\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925491 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925550 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925582 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/56f6dd56-8bbf-4610-b48c-898cda2e2aab-lock\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: E1127 11:55:23.925761 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 11:55:23 crc kubenswrapper[4796]: E1127 11:55:23.925794 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 11:55:23 crc kubenswrapper[4796]: E1127 11:55:23.925843 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift podName:56f6dd56-8bbf-4610-b48c-898cda2e2aab nodeName:}" failed. No retries permitted until 2025-11-27 11:55:24.425828151 +0000 UTC m=+1841.944147169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift") pod "swift-storage-0" (UID: "56f6dd56-8bbf-4610-b48c-898cda2e2aab") : configmap "swift-ring-files" not found Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925976 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.925979 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/56f6dd56-8bbf-4610-b48c-898cda2e2aab-cache\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.926105 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/56f6dd56-8bbf-4610-b48c-898cda2e2aab-lock\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.950218 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:23 crc kubenswrapper[4796]: I1127 11:55:23.954486 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdbbx\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-kube-api-access-wdbbx\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027530 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/469ea57e-e812-49e4-a3f3-947bf53deb22-config\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027598 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kllx9\" (UniqueName: \"kubernetes.io/projected/469ea57e-e812-49e4-a3f3-947bf53deb22-kube-api-access-kllx9\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027649 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027694 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027738 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/469ea57e-e812-49e4-a3f3-947bf53deb22-scripts\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027753 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.027916 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/469ea57e-e812-49e4-a3f3-947bf53deb22-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129310 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/469ea57e-e812-49e4-a3f3-947bf53deb22-config\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129374 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kllx9\" (UniqueName: \"kubernetes.io/projected/469ea57e-e812-49e4-a3f3-947bf53deb22-kube-api-access-kllx9\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129402 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129448 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129503 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/469ea57e-e812-49e4-a3f3-947bf53deb22-scripts\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129525 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.129577 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/469ea57e-e812-49e4-a3f3-947bf53deb22-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.130135 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/469ea57e-e812-49e4-a3f3-947bf53deb22-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.130816 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/469ea57e-e812-49e4-a3f3-947bf53deb22-scripts\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.131915 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/469ea57e-e812-49e4-a3f3-947bf53deb22-config\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.132318 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.134500 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.136046 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/469ea57e-e812-49e4-a3f3-947bf53deb22-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.154927 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kllx9\" (UniqueName: \"kubernetes.io/projected/469ea57e-e812-49e4-a3f3-947bf53deb22-kube-api-access-kllx9\") pod \"ovn-northd-0\" (UID: \"469ea57e-e812-49e4-a3f3-947bf53deb22\") " pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.434228 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:24 crc kubenswrapper[4796]: E1127 11:55:24.434446 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 11:55:24 crc kubenswrapper[4796]: E1127 11:55:24.434475 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 11:55:24 crc kubenswrapper[4796]: E1127 11:55:24.434539 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift podName:56f6dd56-8bbf-4610-b48c-898cda2e2aab nodeName:}" failed. No retries permitted until 2025-11-27 11:55:25.434517531 +0000 UTC m=+1842.952836449 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift") pod "swift-storage-0" (UID: "56f6dd56-8bbf-4610-b48c-898cda2e2aab") : configmap "swift-ring-files" not found Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.452133 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.504857 4796 generic.go:334] "Generic (PLEG): container finished" podID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerID="e308b50574cab3eb6af00d23caf72872d8887a642827246b96ecfeec2233706e" exitCode=0 Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.504942 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" event={"ID":"00d9b05c-b63f-451b-bda5-1451a05ee320","Type":"ContainerDied","Data":"e308b50574cab3eb6af00d23caf72872d8887a642827246b96ecfeec2233706e"} Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.510748 4796 generic.go:334] "Generic (PLEG): container finished" podID="d7630686-df2d-484c-81d3-769308420b18" containerID="d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e" exitCode=0 Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.510803 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" event={"ID":"d7630686-df2d-484c-81d3-769308420b18","Type":"ContainerDied","Data":"d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e"} Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.851429 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.920633 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.942978 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xts4b\" (UniqueName: \"kubernetes.io/projected/00d9b05c-b63f-451b-bda5-1451a05ee320-kube-api-access-xts4b\") pod \"00d9b05c-b63f-451b-bda5-1451a05ee320\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.943083 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-config\") pod \"00d9b05c-b63f-451b-bda5-1451a05ee320\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.943152 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-dns-svc\") pod \"00d9b05c-b63f-451b-bda5-1451a05ee320\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.943311 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-ovsdbserver-sb\") pod \"00d9b05c-b63f-451b-bda5-1451a05ee320\" (UID: \"00d9b05c-b63f-451b-bda5-1451a05ee320\") " Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.954070 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d9b05c-b63f-451b-bda5-1451a05ee320-kube-api-access-xts4b" (OuterVolumeSpecName: "kube-api-access-xts4b") pod "00d9b05c-b63f-451b-bda5-1451a05ee320" (UID: "00d9b05c-b63f-451b-bda5-1451a05ee320"). InnerVolumeSpecName "kube-api-access-xts4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:24 crc kubenswrapper[4796]: I1127 11:55:24.999333 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00d9b05c-b63f-451b-bda5-1451a05ee320" (UID: "00d9b05c-b63f-451b-bda5-1451a05ee320"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:25 crc kubenswrapper[4796]: W1127 11:55:25.015992 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod469ea57e_e812_49e4_a3f3_947bf53deb22.slice/crio-90b4fcf043195154179da206da55ff589f7cc9f96ece87a160343dd0fa6d4ce1 WatchSource:0}: Error finding container 90b4fcf043195154179da206da55ff589f7cc9f96ece87a160343dd0fa6d4ce1: Status 404 returned error can't find the container with id 90b4fcf043195154179da206da55ff589f7cc9f96ece87a160343dd0fa6d4ce1 Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.017192 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.018453 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00d9b05c-b63f-451b-bda5-1451a05ee320" (UID: "00d9b05c-b63f-451b-bda5-1451a05ee320"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.019011 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.036107 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-config" (OuterVolumeSpecName: "config") pod "00d9b05c-b63f-451b-bda5-1451a05ee320" (UID: "00d9b05c-b63f-451b-bda5-1451a05ee320"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.045855 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xts4b\" (UniqueName: \"kubernetes.io/projected/00d9b05c-b63f-451b-bda5-1451a05ee320-kube-api-access-xts4b\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.045896 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.045908 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.045919 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00d9b05c-b63f-451b-bda5-1451a05ee320-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.478645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:25 crc kubenswrapper[4796]: E1127 11:55:25.478823 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 11:55:25 crc kubenswrapper[4796]: E1127 11:55:25.478837 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 11:55:25 crc kubenswrapper[4796]: E1127 11:55:25.478888 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift podName:56f6dd56-8bbf-4610-b48c-898cda2e2aab nodeName:}" failed. No retries permitted until 2025-11-27 11:55:27.478863401 +0000 UTC m=+1844.997182319 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift") pod "swift-storage-0" (UID: "56f6dd56-8bbf-4610-b48c-898cda2e2aab") : configmap "swift-ring-files" not found Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.519756 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" event={"ID":"00d9b05c-b63f-451b-bda5-1451a05ee320","Type":"ContainerDied","Data":"a793f1bffcd11ef856df7a016163fc3d4d37f69a537ccb570822d75c1179ce66"} Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.519816 4796 scope.go:117] "RemoveContainer" containerID="e308b50574cab3eb6af00d23caf72872d8887a642827246b96ecfeec2233706e" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.519955 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.525098 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" event={"ID":"d7630686-df2d-484c-81d3-769308420b18","Type":"ContainerStarted","Data":"4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e"} Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.525611 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.526669 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"469ea57e-e812-49e4-a3f3-947bf53deb22","Type":"ContainerStarted","Data":"90b4fcf043195154179da206da55ff589f7cc9f96ece87a160343dd0fa6d4ce1"} Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.550262 4796 scope.go:117] "RemoveContainer" containerID="796ff09375965554e3face8ac11beca006bce3c45eec55671118b5d267a19796" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.552290 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" podStartSLOduration=3.552257419 podStartE2EDuration="3.552257419s" podCreationTimestamp="2025-11-27 11:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:55:25.546376262 +0000 UTC m=+1843.064695210" watchObservedRunningTime="2025-11-27 11:55:25.552257419 +0000 UTC m=+1843.070576347" Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.564286 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9c62v"] Nov 27 11:55:25 crc kubenswrapper[4796]: I1127 11:55:25.582016 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-9c62v"] Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.543013 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"469ea57e-e812-49e4-a3f3-947bf53deb22","Type":"ContainerStarted","Data":"b353aeaa559f0f099286061615283144a32891212587349047c86010ba39b487"} Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.543736 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"469ea57e-e812-49e4-a3f3-947bf53deb22","Type":"ContainerStarted","Data":"3c535bd15c7b9b2b9053ac836feb5f19706613d39fa8c9eec6ca34ce9b8d9d63"} Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.543772 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.568679 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jw9v8"] Nov 27 11:55:27 crc kubenswrapper[4796]: E1127 11:55:27.569308 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="dnsmasq-dns" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.569336 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="dnsmasq-dns" Nov 27 11:55:27 crc kubenswrapper[4796]: E1127 11:55:27.569366 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="init" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.569379 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="init" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.569703 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="dnsmasq-dns" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.570677 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.573058 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.573300 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.573364 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.573439 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.858721813 podStartE2EDuration="4.573422327s" podCreationTimestamp="2025-11-27 11:55:23 +0000 UTC" firstStartedPulling="2025-11-27 11:55:25.017974595 +0000 UTC m=+1842.536293503" lastFinishedPulling="2025-11-27 11:55:26.732675099 +0000 UTC m=+1844.250994017" observedRunningTime="2025-11-27 11:55:27.561401676 +0000 UTC m=+1845.079720594" watchObservedRunningTime="2025-11-27 11:55:27.573422327 +0000 UTC m=+1845.091741245" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.578513 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/085a80b3-9f90-480a-8e26-fda7fb500c64-etc-swift\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.578595 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-ring-data-devices\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.578621 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-dispersionconf\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.579045 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-swiftconf\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.579100 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-scripts\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.579174 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.579211 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-combined-ca-bundle\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.579419 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7mnl\" (UniqueName: \"kubernetes.io/projected/085a80b3-9f90-480a-8e26-fda7fb500c64-kube-api-access-g7mnl\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: E1127 11:55:27.579521 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 11:55:27 crc kubenswrapper[4796]: E1127 11:55:27.579541 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 11:55:27 crc kubenswrapper[4796]: E1127 11:55:27.579577 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift podName:56f6dd56-8bbf-4610-b48c-898cda2e2aab nodeName:}" failed. No retries permitted until 2025-11-27 11:55:31.579564191 +0000 UTC m=+1849.097883109 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift") pod "swift-storage-0" (UID: "56f6dd56-8bbf-4610-b48c-898cda2e2aab") : configmap "swift-ring-files" not found Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.588339 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" path="/var/lib/kubelet/pods/00d9b05c-b63f-451b-bda5-1451a05ee320/volumes" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.601363 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jw9v8"] Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.681507 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-ring-data-devices\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.681598 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-dispersionconf\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.681679 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-swiftconf\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.681731 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-scripts\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.681849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-combined-ca-bundle\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.681897 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7mnl\" (UniqueName: \"kubernetes.io/projected/085a80b3-9f90-480a-8e26-fda7fb500c64-kube-api-access-g7mnl\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.682005 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/085a80b3-9f90-480a-8e26-fda7fb500c64-etc-swift\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.682798 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/085a80b3-9f90-480a-8e26-fda7fb500c64-etc-swift\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.684218 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-ring-data-devices\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.686630 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-scripts\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.689343 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-dispersionconf\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.689921 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-swiftconf\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.689932 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-combined-ca-bundle\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.706641 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7mnl\" (UniqueName: \"kubernetes.io/projected/085a80b3-9f90-480a-8e26-fda7fb500c64-kube-api-access-g7mnl\") pod \"swift-ring-rebalance-jw9v8\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:27 crc kubenswrapper[4796]: I1127 11:55:27.887978 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:28 crc kubenswrapper[4796]: I1127 11:55:28.353196 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jw9v8"] Nov 27 11:55:28 crc kubenswrapper[4796]: W1127 11:55:28.360805 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod085a80b3_9f90_480a_8e26_fda7fb500c64.slice/crio-ce47ce5ae279ea3e18ef8644c8d62a361e5d3175fd66a57f5051a11b68efbc7e WatchSource:0}: Error finding container ce47ce5ae279ea3e18ef8644c8d62a361e5d3175fd66a57f5051a11b68efbc7e: Status 404 returned error can't find the container with id ce47ce5ae279ea3e18ef8644c8d62a361e5d3175fd66a57f5051a11b68efbc7e Nov 27 11:55:28 crc kubenswrapper[4796]: I1127 11:55:28.553236 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw9v8" event={"ID":"085a80b3-9f90-480a-8e26-fda7fb500c64","Type":"ContainerStarted","Data":"ce47ce5ae279ea3e18ef8644c8d62a361e5d3175fd66a57f5051a11b68efbc7e"} Nov 27 11:55:28 crc kubenswrapper[4796]: I1127 11:55:28.974809 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 27 11:55:28 crc kubenswrapper[4796]: I1127 11:55:28.974883 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 27 11:55:29 crc kubenswrapper[4796]: I1127 11:55:29.099222 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 27 11:55:29 crc kubenswrapper[4796]: I1127 11:55:29.641954 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 27 11:55:29 crc kubenswrapper[4796]: I1127 11:55:29.685930 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bc7876d45-9c62v" podUID="00d9b05c-b63f-451b-bda5-1451a05ee320" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.153537 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9d19-account-create-update-f7228"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.156790 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.161598 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.172543 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9d19-account-create-update-f7228"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.232513 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-7shzf"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.233713 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.240853 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7shzf"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.318671 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.318957 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.337324 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp8fs\" (UniqueName: \"kubernetes.io/projected/6cab3a5d-5bee-478c-9597-6742f2357e86-kube-api-access-sp8fs\") pod \"keystone-9d19-account-create-update-f7228\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.337526 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp2bn\" (UniqueName: \"kubernetes.io/projected/7db284c6-3140-4359-b58b-e3868608590a-kube-api-access-sp2bn\") pod \"keystone-db-create-7shzf\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.337576 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cab3a5d-5bee-478c-9597-6742f2357e86-operator-scripts\") pod \"keystone-9d19-account-create-update-f7228\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.337622 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db284c6-3140-4359-b58b-e3868608590a-operator-scripts\") pod \"keystone-db-create-7shzf\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.392958 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-hxhjx"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.394338 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.399788 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hxhjx"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.415189 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.439220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db284c6-3140-4359-b58b-e3868608590a-operator-scripts\") pod \"keystone-db-create-7shzf\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.439372 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp8fs\" (UniqueName: \"kubernetes.io/projected/6cab3a5d-5bee-478c-9597-6742f2357e86-kube-api-access-sp8fs\") pod \"keystone-9d19-account-create-update-f7228\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.439476 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp2bn\" (UniqueName: \"kubernetes.io/projected/7db284c6-3140-4359-b58b-e3868608590a-kube-api-access-sp2bn\") pod \"keystone-db-create-7shzf\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.439507 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cab3a5d-5bee-478c-9597-6742f2357e86-operator-scripts\") pod \"keystone-9d19-account-create-update-f7228\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.440172 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cab3a5d-5bee-478c-9597-6742f2357e86-operator-scripts\") pod \"keystone-9d19-account-create-update-f7228\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.440735 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db284c6-3140-4359-b58b-e3868608590a-operator-scripts\") pod \"keystone-db-create-7shzf\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.459857 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp2bn\" (UniqueName: \"kubernetes.io/projected/7db284c6-3140-4359-b58b-e3868608590a-kube-api-access-sp2bn\") pod \"keystone-db-create-7shzf\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.462751 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp8fs\" (UniqueName: \"kubernetes.io/projected/6cab3a5d-5bee-478c-9597-6742f2357e86-kube-api-access-sp8fs\") pod \"keystone-9d19-account-create-update-f7228\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.479763 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.507222 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-cbac-account-create-update-jn7bg"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.508407 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.513098 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.529695 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cbac-account-create-update-jn7bg"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.541116 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407847a5-70c0-43f0-b86f-3b5d6e64ff22-operator-scripts\") pod \"placement-db-create-hxhjx\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.541187 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh2mk\" (UniqueName: \"kubernetes.io/projected/407847a5-70c0-43f0-b86f-3b5d6e64ff22-kube-api-access-wh2mk\") pod \"placement-db-create-hxhjx\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.557154 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.643053 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fff7060c-0cf3-49e4-811a-aeaf656601a0-operator-scripts\") pod \"placement-cbac-account-create-update-jn7bg\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.643118 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407847a5-70c0-43f0-b86f-3b5d6e64ff22-operator-scripts\") pod \"placement-db-create-hxhjx\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.643168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khsnv\" (UniqueName: \"kubernetes.io/projected/fff7060c-0cf3-49e4-811a-aeaf656601a0-kube-api-access-khsnv\") pod \"placement-cbac-account-create-update-jn7bg\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.643192 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh2mk\" (UniqueName: \"kubernetes.io/projected/407847a5-70c0-43f0-b86f-3b5d6e64ff22-kube-api-access-wh2mk\") pod \"placement-db-create-hxhjx\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.644187 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407847a5-70c0-43f0-b86f-3b5d6e64ff22-operator-scripts\") pod \"placement-db-create-hxhjx\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.662972 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh2mk\" (UniqueName: \"kubernetes.io/projected/407847a5-70c0-43f0-b86f-3b5d6e64ff22-kube-api-access-wh2mk\") pod \"placement-db-create-hxhjx\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.671605 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.727479 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.741228 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ngbvt"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.742644 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.744137 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fff7060c-0cf3-49e4-811a-aeaf656601a0-operator-scripts\") pod \"placement-cbac-account-create-update-jn7bg\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.744234 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khsnv\" (UniqueName: \"kubernetes.io/projected/fff7060c-0cf3-49e4-811a-aeaf656601a0-kube-api-access-khsnv\") pod \"placement-cbac-account-create-update-jn7bg\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.746704 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fff7060c-0cf3-49e4-811a-aeaf656601a0-operator-scripts\") pod \"placement-cbac-account-create-update-jn7bg\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.753168 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ngbvt"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.765926 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khsnv\" (UniqueName: \"kubernetes.io/projected/fff7060c-0cf3-49e4-811a-aeaf656601a0-kube-api-access-khsnv\") pod \"placement-cbac-account-create-update-jn7bg\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.846847 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.848046 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-operator-scripts\") pod \"glance-db-create-ngbvt\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.848149 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r2dp\" (UniqueName: \"kubernetes.io/projected/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-kube-api-access-2r2dp\") pod \"glance-db-create-ngbvt\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.850165 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b679-account-create-update-pbbpm"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.851585 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.853979 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.861986 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b679-account-create-update-pbbpm"] Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.949986 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7b76a35-a5ea-4a21-9858-a48f47b6bead-operator-scripts\") pod \"glance-b679-account-create-update-pbbpm\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.950068 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-operator-scripts\") pod \"glance-db-create-ngbvt\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.950105 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r2dp\" (UniqueName: \"kubernetes.io/projected/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-kube-api-access-2r2dp\") pod \"glance-db-create-ngbvt\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.950126 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prtxj\" (UniqueName: \"kubernetes.io/projected/c7b76a35-a5ea-4a21-9858-a48f47b6bead-kube-api-access-prtxj\") pod \"glance-b679-account-create-update-pbbpm\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.950743 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-operator-scripts\") pod \"glance-db-create-ngbvt\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:30 crc kubenswrapper[4796]: I1127 11:55:30.979976 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r2dp\" (UniqueName: \"kubernetes.io/projected/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-kube-api-access-2r2dp\") pod \"glance-db-create-ngbvt\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.052439 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7b76a35-a5ea-4a21-9858-a48f47b6bead-operator-scripts\") pod \"glance-b679-account-create-update-pbbpm\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.052625 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prtxj\" (UniqueName: \"kubernetes.io/projected/c7b76a35-a5ea-4a21-9858-a48f47b6bead-kube-api-access-prtxj\") pod \"glance-b679-account-create-update-pbbpm\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.053243 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7b76a35-a5ea-4a21-9858-a48f47b6bead-operator-scripts\") pod \"glance-b679-account-create-update-pbbpm\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.074544 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prtxj\" (UniqueName: \"kubernetes.io/projected/c7b76a35-a5ea-4a21-9858-a48f47b6bead-kube-api-access-prtxj\") pod \"glance-b679-account-create-update-pbbpm\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.105815 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.174683 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:31 crc kubenswrapper[4796]: I1127 11:55:31.662817 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:31 crc kubenswrapper[4796]: E1127 11:55:31.665747 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 11:55:31 crc kubenswrapper[4796]: E1127 11:55:31.665772 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 11:55:31 crc kubenswrapper[4796]: E1127 11:55:31.665805 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift podName:56f6dd56-8bbf-4610-b48c-898cda2e2aab nodeName:}" failed. No retries permitted until 2025-11-27 11:55:39.665792548 +0000 UTC m=+1857.184111466 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift") pod "swift-storage-0" (UID: "56f6dd56-8bbf-4610-b48c-898cda2e2aab") : configmap "swift-ring-files" not found Nov 27 11:55:32 crc kubenswrapper[4796]: I1127 11:55:32.893532 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cbac-account-create-update-jn7bg"] Nov 27 11:55:32 crc kubenswrapper[4796]: I1127 11:55:32.952721 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.049956 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnwth"] Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.050205 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-mnwth" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerName="dnsmasq-dns" containerID="cri-o://3ad28e41a17204998b025ee5336700e2f49ffd055f36804a31829a269e5f49dc" gracePeriod=10 Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.087835 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hxhjx"] Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.121621 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7shzf"] Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.213166 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b679-account-create-update-pbbpm"] Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.226084 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ngbvt"] Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.260358 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9d19-account-create-update-f7228"] Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.592983 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7shzf" event={"ID":"7db284c6-3140-4359-b58b-e3868608590a","Type":"ContainerStarted","Data":"51f943c70a0403d361fa6f7378b63cd404a0c9a667f2c29329c5fe5612b031cc"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.593329 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7shzf" event={"ID":"7db284c6-3140-4359-b58b-e3868608590a","Type":"ContainerStarted","Data":"d087bd1a831b7deafea01554174e0f752e25100a82259c25ce3c1a6eac0440c1"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.597528 4796 generic.go:334] "Generic (PLEG): container finished" podID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerID="3ad28e41a17204998b025ee5336700e2f49ffd055f36804a31829a269e5f49dc" exitCode=0 Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.597600 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnwth" event={"ID":"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940","Type":"ContainerDied","Data":"3ad28e41a17204998b025ee5336700e2f49ffd055f36804a31829a269e5f49dc"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.601229 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hxhjx" event={"ID":"407847a5-70c0-43f0-b86f-3b5d6e64ff22","Type":"ContainerStarted","Data":"0c021bce9cd8479a5fb869b93b1bba82be398e2a123aef602dd5993947771b65"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.601287 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hxhjx" event={"ID":"407847a5-70c0-43f0-b86f-3b5d6e64ff22","Type":"ContainerStarted","Data":"dfbb92b31683a4147c619817e484df72e712af1b5d9e2f7fe07383e70babf21c"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.615607 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ngbvt" event={"ID":"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6","Type":"ContainerStarted","Data":"3a365ff88e03082ed880b00386d400b9bc2970ac2bab6e65494c89318ce4c010"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.615666 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ngbvt" event={"ID":"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6","Type":"ContainerStarted","Data":"0d74105f6ebec35cfa03812ef539cd263c690bd7ea878f262f9854d9a885cf11"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.617141 4796 generic.go:334] "Generic (PLEG): container finished" podID="fff7060c-0cf3-49e4-811a-aeaf656601a0" containerID="5217d4919a1f79de0a525fac7d36476a70e4cbf6ea33327718d60edb741d9991" exitCode=0 Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.617479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cbac-account-create-update-jn7bg" event={"ID":"fff7060c-0cf3-49e4-811a-aeaf656601a0","Type":"ContainerDied","Data":"5217d4919a1f79de0a525fac7d36476a70e4cbf6ea33327718d60edb741d9991"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.617511 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cbac-account-create-update-jn7bg" event={"ID":"fff7060c-0cf3-49e4-811a-aeaf656601a0","Type":"ContainerStarted","Data":"0cd6e56f7afe8132c1fc50bf49f0c98779e3360a0a0e9259d1c1ed0a72925835"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.619389 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9d19-account-create-update-f7228" event={"ID":"6cab3a5d-5bee-478c-9597-6742f2357e86","Type":"ContainerStarted","Data":"72504b6c88ea44cf28e7a64fb9ac76c726e04b938d99f1c9bb579376098afa80"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.619414 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9d19-account-create-update-f7228" event={"ID":"6cab3a5d-5bee-478c-9597-6742f2357e86","Type":"ContainerStarted","Data":"91ac9737d6bd6fe8e9f6c30e01f4b0101847a851afd3f0e0a71fc651e3e42673"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.621904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw9v8" event={"ID":"085a80b3-9f90-480a-8e26-fda7fb500c64","Type":"ContainerStarted","Data":"05ae02bc1d108b3791617c81afff1dd5061f0c8718f4915c0e413b3a5c593d6c"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.624322 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b679-account-create-update-pbbpm" event={"ID":"c7b76a35-a5ea-4a21-9858-a48f47b6bead","Type":"ContainerStarted","Data":"1032d80608d2b61125c35d5d4f01c35936ecfc6db90403990e6b0e05b996472a"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.624372 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b679-account-create-update-pbbpm" event={"ID":"c7b76a35-a5ea-4a21-9858-a48f47b6bead","Type":"ContainerStarted","Data":"b89de8e3902cb335607ddcb9fde137dbb75eccd370f1d1ab64effaf093446817"} Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.761298 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-ngbvt" podStartSLOduration=3.761258479 podStartE2EDuration="3.761258479s" podCreationTimestamp="2025-11-27 11:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:55:33.757485199 +0000 UTC m=+1851.275804107" watchObservedRunningTime="2025-11-27 11:55:33.761258479 +0000 UTC m=+1851.279577397" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.776838 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-7shzf" podStartSLOduration=3.776820545 podStartE2EDuration="3.776820545s" podCreationTimestamp="2025-11-27 11:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:55:33.769200551 +0000 UTC m=+1851.287519459" watchObservedRunningTime="2025-11-27 11:55:33.776820545 +0000 UTC m=+1851.295139463" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.794424 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-hxhjx" podStartSLOduration=3.7944022029999998 podStartE2EDuration="3.794402203s" podCreationTimestamp="2025-11-27 11:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:55:33.79087711 +0000 UTC m=+1851.309196028" watchObservedRunningTime="2025-11-27 11:55:33.794402203 +0000 UTC m=+1851.312721121" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.821727 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9d19-account-create-update-f7228" podStartSLOduration=3.821707542 podStartE2EDuration="3.821707542s" podCreationTimestamp="2025-11-27 11:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:55:33.804942945 +0000 UTC m=+1851.323261853" watchObservedRunningTime="2025-11-27 11:55:33.821707542 +0000 UTC m=+1851.340026460" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.824383 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-b679-account-create-update-pbbpm" podStartSLOduration=3.824373723 podStartE2EDuration="3.824373723s" podCreationTimestamp="2025-11-27 11:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:55:33.818026724 +0000 UTC m=+1851.336345642" watchObservedRunningTime="2025-11-27 11:55:33.824373723 +0000 UTC m=+1851.342692641" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.841719 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-jw9v8" podStartSLOduration=2.769149911 podStartE2EDuration="6.841700535s" podCreationTimestamp="2025-11-27 11:55:27 +0000 UTC" firstStartedPulling="2025-11-27 11:55:28.363992957 +0000 UTC m=+1845.882311875" lastFinishedPulling="2025-11-27 11:55:32.436543581 +0000 UTC m=+1849.954862499" observedRunningTime="2025-11-27 11:55:33.836659081 +0000 UTC m=+1851.354978019" watchObservedRunningTime="2025-11-27 11:55:33.841700535 +0000 UTC m=+1851.360019453" Nov 27 11:55:33 crc kubenswrapper[4796]: I1127 11:55:33.993847 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.126136 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-sb\") pod \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.126404 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-config\") pod \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.126761 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-dns-svc\") pod \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.126898 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-nb\") pod \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.126961 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b7x4\" (UniqueName: \"kubernetes.io/projected/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-kube-api-access-2b7x4\") pod \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\" (UID: \"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940\") " Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.143753 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-kube-api-access-2b7x4" (OuterVolumeSpecName: "kube-api-access-2b7x4") pod "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" (UID: "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940"). InnerVolumeSpecName "kube-api-access-2b7x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.182506 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" (UID: "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.183314 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-config" (OuterVolumeSpecName: "config") pod "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" (UID: "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.193764 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" (UID: "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.205400 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" (UID: "fb13cfe9-24c9-41ef-a71c-6f8c63b7f940"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.229184 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.229232 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.229243 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.229279 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b7x4\" (UniqueName: \"kubernetes.io/projected/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-kube-api-access-2b7x4\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.229292 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.636098 4796 generic.go:334] "Generic (PLEG): container finished" podID="407847a5-70c0-43f0-b86f-3b5d6e64ff22" containerID="0c021bce9cd8479a5fb869b93b1bba82be398e2a123aef602dd5993947771b65" exitCode=0 Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.636182 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hxhjx" event={"ID":"407847a5-70c0-43f0-b86f-3b5d6e64ff22","Type":"ContainerDied","Data":"0c021bce9cd8479a5fb869b93b1bba82be398e2a123aef602dd5993947771b65"} Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.638306 4796 generic.go:334] "Generic (PLEG): container finished" podID="25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" containerID="3a365ff88e03082ed880b00386d400b9bc2970ac2bab6e65494c89318ce4c010" exitCode=0 Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.638391 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ngbvt" event={"ID":"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6","Type":"ContainerDied","Data":"3a365ff88e03082ed880b00386d400b9bc2970ac2bab6e65494c89318ce4c010"} Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.640311 4796 generic.go:334] "Generic (PLEG): container finished" podID="6cab3a5d-5bee-478c-9597-6742f2357e86" containerID="72504b6c88ea44cf28e7a64fb9ac76c726e04b938d99f1c9bb579376098afa80" exitCode=0 Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.640405 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9d19-account-create-update-f7228" event={"ID":"6cab3a5d-5bee-478c-9597-6742f2357e86","Type":"ContainerDied","Data":"72504b6c88ea44cf28e7a64fb9ac76c726e04b938d99f1c9bb579376098afa80"} Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.641931 4796 generic.go:334] "Generic (PLEG): container finished" podID="c7b76a35-a5ea-4a21-9858-a48f47b6bead" containerID="1032d80608d2b61125c35d5d4f01c35936ecfc6db90403990e6b0e05b996472a" exitCode=0 Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.641982 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b679-account-create-update-pbbpm" event={"ID":"c7b76a35-a5ea-4a21-9858-a48f47b6bead","Type":"ContainerDied","Data":"1032d80608d2b61125c35d5d4f01c35936ecfc6db90403990e6b0e05b996472a"} Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.644207 4796 generic.go:334] "Generic (PLEG): container finished" podID="7db284c6-3140-4359-b58b-e3868608590a" containerID="51f943c70a0403d361fa6f7378b63cd404a0c9a667f2c29329c5fe5612b031cc" exitCode=0 Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.644319 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7shzf" event={"ID":"7db284c6-3140-4359-b58b-e3868608590a","Type":"ContainerDied","Data":"51f943c70a0403d361fa6f7378b63cd404a0c9a667f2c29329c5fe5612b031cc"} Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.646540 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnwth" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.656690 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnwth" event={"ID":"fb13cfe9-24c9-41ef-a71c-6f8c63b7f940","Type":"ContainerDied","Data":"21994199e6d9e358a1c0153515b8a6ba5e7f2331027796e7f751e2963caad517"} Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.656782 4796 scope.go:117] "RemoveContainer" containerID="3ad28e41a17204998b025ee5336700e2f49ffd055f36804a31829a269e5f49dc" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.691466 4796 scope.go:117] "RemoveContainer" containerID="b26b517c154e56b5dd4c04cc5b2a15216138edf9e963ef2910f6bf75c2345426" Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.766985 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnwth"] Nov 27 11:55:34 crc kubenswrapper[4796]: I1127 11:55:34.773205 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnwth"] Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.067064 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.149578 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khsnv\" (UniqueName: \"kubernetes.io/projected/fff7060c-0cf3-49e4-811a-aeaf656601a0-kube-api-access-khsnv\") pod \"fff7060c-0cf3-49e4-811a-aeaf656601a0\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.149877 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fff7060c-0cf3-49e4-811a-aeaf656601a0-operator-scripts\") pod \"fff7060c-0cf3-49e4-811a-aeaf656601a0\" (UID: \"fff7060c-0cf3-49e4-811a-aeaf656601a0\") " Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.155074 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fff7060c-0cf3-49e4-811a-aeaf656601a0-kube-api-access-khsnv" (OuterVolumeSpecName: "kube-api-access-khsnv") pod "fff7060c-0cf3-49e4-811a-aeaf656601a0" (UID: "fff7060c-0cf3-49e4-811a-aeaf656601a0"). InnerVolumeSpecName "kube-api-access-khsnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.160879 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fff7060c-0cf3-49e4-811a-aeaf656601a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fff7060c-0cf3-49e4-811a-aeaf656601a0" (UID: "fff7060c-0cf3-49e4-811a-aeaf656601a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.252193 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fff7060c-0cf3-49e4-811a-aeaf656601a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.252241 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khsnv\" (UniqueName: \"kubernetes.io/projected/fff7060c-0cf3-49e4-811a-aeaf656601a0-kube-api-access-khsnv\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.579898 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" path="/var/lib/kubelet/pods/fb13cfe9-24c9-41ef-a71c-6f8c63b7f940/volumes" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.656970 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cbac-account-create-update-jn7bg" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.657926 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cbac-account-create-update-jn7bg" event={"ID":"fff7060c-0cf3-49e4-811a-aeaf656601a0","Type":"ContainerDied","Data":"0cd6e56f7afe8132c1fc50bf49f0c98779e3360a0a0e9259d1c1ed0a72925835"} Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.657962 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cd6e56f7afe8132c1fc50bf49f0c98779e3360a0a0e9259d1c1ed0a72925835" Nov 27 11:55:35 crc kubenswrapper[4796]: I1127 11:55:35.979262 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.068015 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp2bn\" (UniqueName: \"kubernetes.io/projected/7db284c6-3140-4359-b58b-e3868608590a-kube-api-access-sp2bn\") pod \"7db284c6-3140-4359-b58b-e3868608590a\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.068104 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db284c6-3140-4359-b58b-e3868608590a-operator-scripts\") pod \"7db284c6-3140-4359-b58b-e3868608590a\" (UID: \"7db284c6-3140-4359-b58b-e3868608590a\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.068634 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db284c6-3140-4359-b58b-e3868608590a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7db284c6-3140-4359-b58b-e3868608590a" (UID: "7db284c6-3140-4359-b58b-e3868608590a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.073798 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db284c6-3140-4359-b58b-e3868608590a-kube-api-access-sp2bn" (OuterVolumeSpecName: "kube-api-access-sp2bn") pod "7db284c6-3140-4359-b58b-e3868608590a" (UID: "7db284c6-3140-4359-b58b-e3868608590a"). InnerVolumeSpecName "kube-api-access-sp2bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.169940 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp2bn\" (UniqueName: \"kubernetes.io/projected/7db284c6-3140-4359-b58b-e3868608590a-kube-api-access-sp2bn\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.170509 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db284c6-3140-4359-b58b-e3868608590a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.199800 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.207787 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.275193 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cab3a5d-5bee-478c-9597-6742f2357e86-operator-scripts\") pod \"6cab3a5d-5bee-478c-9597-6742f2357e86\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.275461 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh2mk\" (UniqueName: \"kubernetes.io/projected/407847a5-70c0-43f0-b86f-3b5d6e64ff22-kube-api-access-wh2mk\") pod \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.275501 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407847a5-70c0-43f0-b86f-3b5d6e64ff22-operator-scripts\") pod \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\" (UID: \"407847a5-70c0-43f0-b86f-3b5d6e64ff22\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.275544 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp8fs\" (UniqueName: \"kubernetes.io/projected/6cab3a5d-5bee-478c-9597-6742f2357e86-kube-api-access-sp8fs\") pod \"6cab3a5d-5bee-478c-9597-6742f2357e86\" (UID: \"6cab3a5d-5bee-478c-9597-6742f2357e86\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.277035 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cab3a5d-5bee-478c-9597-6742f2357e86-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6cab3a5d-5bee-478c-9597-6742f2357e86" (UID: "6cab3a5d-5bee-478c-9597-6742f2357e86"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.277410 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407847a5-70c0-43f0-b86f-3b5d6e64ff22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "407847a5-70c0-43f0-b86f-3b5d6e64ff22" (UID: "407847a5-70c0-43f0-b86f-3b5d6e64ff22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.281496 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cab3a5d-5bee-478c-9597-6742f2357e86-kube-api-access-sp8fs" (OuterVolumeSpecName: "kube-api-access-sp8fs") pod "6cab3a5d-5bee-478c-9597-6742f2357e86" (UID: "6cab3a5d-5bee-478c-9597-6742f2357e86"). InnerVolumeSpecName "kube-api-access-sp8fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.298520 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/407847a5-70c0-43f0-b86f-3b5d6e64ff22-kube-api-access-wh2mk" (OuterVolumeSpecName: "kube-api-access-wh2mk") pod "407847a5-70c0-43f0-b86f-3b5d6e64ff22" (UID: "407847a5-70c0-43f0-b86f-3b5d6e64ff22"). InnerVolumeSpecName "kube-api-access-wh2mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.381521 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh2mk\" (UniqueName: \"kubernetes.io/projected/407847a5-70c0-43f0-b86f-3b5d6e64ff22-kube-api-access-wh2mk\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.381554 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407847a5-70c0-43f0-b86f-3b5d6e64ff22-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.381563 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp8fs\" (UniqueName: \"kubernetes.io/projected/6cab3a5d-5bee-478c-9597-6742f2357e86-kube-api-access-sp8fs\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.381573 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cab3a5d-5bee-478c-9597-6742f2357e86-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.383859 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.409896 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.482653 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prtxj\" (UniqueName: \"kubernetes.io/projected/c7b76a35-a5ea-4a21-9858-a48f47b6bead-kube-api-access-prtxj\") pod \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.482694 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r2dp\" (UniqueName: \"kubernetes.io/projected/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-kube-api-access-2r2dp\") pod \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.482769 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7b76a35-a5ea-4a21-9858-a48f47b6bead-operator-scripts\") pod \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\" (UID: \"c7b76a35-a5ea-4a21-9858-a48f47b6bead\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.482901 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-operator-scripts\") pod \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\" (UID: \"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6\") " Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.483564 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" (UID: "25cf6b67-ab55-4aa8-a2d6-40421ec20aa6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.486524 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7b76a35-a5ea-4a21-9858-a48f47b6bead-kube-api-access-prtxj" (OuterVolumeSpecName: "kube-api-access-prtxj") pod "c7b76a35-a5ea-4a21-9858-a48f47b6bead" (UID: "c7b76a35-a5ea-4a21-9858-a48f47b6bead"). InnerVolumeSpecName "kube-api-access-prtxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.486747 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7b76a35-a5ea-4a21-9858-a48f47b6bead-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7b76a35-a5ea-4a21-9858-a48f47b6bead" (UID: "c7b76a35-a5ea-4a21-9858-a48f47b6bead"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.491417 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-kube-api-access-2r2dp" (OuterVolumeSpecName: "kube-api-access-2r2dp") pod "25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" (UID: "25cf6b67-ab55-4aa8-a2d6-40421ec20aa6"). InnerVolumeSpecName "kube-api-access-2r2dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.585073 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.585108 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prtxj\" (UniqueName: \"kubernetes.io/projected/c7b76a35-a5ea-4a21-9858-a48f47b6bead-kube-api-access-prtxj\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.585120 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r2dp\" (UniqueName: \"kubernetes.io/projected/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6-kube-api-access-2r2dp\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.585130 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7b76a35-a5ea-4a21-9858-a48f47b6bead-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.667252 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hxhjx" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.667239 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hxhjx" event={"ID":"407847a5-70c0-43f0-b86f-3b5d6e64ff22","Type":"ContainerDied","Data":"dfbb92b31683a4147c619817e484df72e712af1b5d9e2f7fe07383e70babf21c"} Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.667481 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfbb92b31683a4147c619817e484df72e712af1b5d9e2f7fe07383e70babf21c" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.669034 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ngbvt" event={"ID":"25cf6b67-ab55-4aa8-a2d6-40421ec20aa6","Type":"ContainerDied","Data":"0d74105f6ebec35cfa03812ef539cd263c690bd7ea878f262f9854d9a885cf11"} Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.669066 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d74105f6ebec35cfa03812ef539cd263c690bd7ea878f262f9854d9a885cf11" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.669129 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ngbvt" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.671925 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9d19-account-create-update-f7228" event={"ID":"6cab3a5d-5bee-478c-9597-6742f2357e86","Type":"ContainerDied","Data":"91ac9737d6bd6fe8e9f6c30e01f4b0101847a851afd3f0e0a71fc651e3e42673"} Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.671970 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91ac9737d6bd6fe8e9f6c30e01f4b0101847a851afd3f0e0a71fc651e3e42673" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.672032 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9d19-account-create-update-f7228" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.673862 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b679-account-create-update-pbbpm" event={"ID":"c7b76a35-a5ea-4a21-9858-a48f47b6bead","Type":"ContainerDied","Data":"b89de8e3902cb335607ddcb9fde137dbb75eccd370f1d1ab64effaf093446817"} Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.673892 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b89de8e3902cb335607ddcb9fde137dbb75eccd370f1d1ab64effaf093446817" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.673932 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b679-account-create-update-pbbpm" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.686231 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7shzf" event={"ID":"7db284c6-3140-4359-b58b-e3868608590a","Type":"ContainerDied","Data":"d087bd1a831b7deafea01554174e0f752e25100a82259c25ce3c1a6eac0440c1"} Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.686312 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d087bd1a831b7deafea01554174e0f752e25100a82259c25ce3c1a6eac0440c1" Nov 27 11:55:36 crc kubenswrapper[4796]: I1127 11:55:36.686402 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7shzf" Nov 27 11:55:37 crc kubenswrapper[4796]: I1127 11:55:37.569239 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:55:37 crc kubenswrapper[4796]: E1127 11:55:37.569690 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:55:39 crc kubenswrapper[4796]: I1127 11:55:39.527363 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 27 11:55:39 crc kubenswrapper[4796]: I1127 11:55:39.735635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:39 crc kubenswrapper[4796]: E1127 11:55:39.736107 4796 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 27 11:55:39 crc kubenswrapper[4796]: E1127 11:55:39.736121 4796 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 27 11:55:39 crc kubenswrapper[4796]: E1127 11:55:39.736162 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift podName:56f6dd56-8bbf-4610-b48c-898cda2e2aab nodeName:}" failed. No retries permitted until 2025-11-27 11:55:55.736148651 +0000 UTC m=+1873.254467569 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift") pod "swift-storage-0" (UID: "56f6dd56-8bbf-4610-b48c-898cda2e2aab") : configmap "swift-ring-files" not found Nov 27 11:55:40 crc kubenswrapper[4796]: I1127 11:55:40.723928 4796 generic.go:334] "Generic (PLEG): container finished" podID="085a80b3-9f90-480a-8e26-fda7fb500c64" containerID="05ae02bc1d108b3791617c81afff1dd5061f0c8718f4915c0e413b3a5c593d6c" exitCode=0 Nov 27 11:55:40 crc kubenswrapper[4796]: I1127 11:55:40.723980 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw9v8" event={"ID":"085a80b3-9f90-480a-8e26-fda7fb500c64","Type":"ContainerDied","Data":"05ae02bc1d108b3791617c81afff1dd5061f0c8718f4915c0e413b3a5c593d6c"} Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.052623 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-rc66k"] Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053006 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerName="dnsmasq-dns" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053029 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerName="dnsmasq-dns" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053042 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407847a5-70c0-43f0-b86f-3b5d6e64ff22" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053052 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="407847a5-70c0-43f0-b86f-3b5d6e64ff22" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053066 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerName="init" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053074 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerName="init" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053099 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cab3a5d-5bee-478c-9597-6742f2357e86" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053107 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cab3a5d-5bee-478c-9597-6742f2357e86" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053121 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053129 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053143 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fff7060c-0cf3-49e4-811a-aeaf656601a0" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053151 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fff7060c-0cf3-49e4-811a-aeaf656601a0" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053169 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7b76a35-a5ea-4a21-9858-a48f47b6bead" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053179 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7b76a35-a5ea-4a21-9858-a48f47b6bead" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: E1127 11:55:41.053192 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db284c6-3140-4359-b58b-e3868608590a" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053200 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db284c6-3140-4359-b58b-e3868608590a" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053413 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="fff7060c-0cf3-49e4-811a-aeaf656601a0" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053452 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053478 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb13cfe9-24c9-41ef-a71c-6f8c63b7f940" containerName="dnsmasq-dns" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053491 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="407847a5-70c0-43f0-b86f-3b5d6e64ff22" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053501 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cab3a5d-5bee-478c-9597-6742f2357e86" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053518 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7b76a35-a5ea-4a21-9858-a48f47b6bead" containerName="mariadb-account-create-update" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.053535 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db284c6-3140-4359-b58b-e3868608590a" containerName="mariadb-database-create" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.054113 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.066309 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.066716 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gbfzq" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.090530 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rc66k"] Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.159202 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-config-data\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.159259 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-combined-ca-bundle\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.159313 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-db-sync-config-data\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.159387 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkr4g\" (UniqueName: \"kubernetes.io/projected/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-kube-api-access-nkr4g\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.260749 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-combined-ca-bundle\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.260798 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-db-sync-config-data\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.260873 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkr4g\" (UniqueName: \"kubernetes.io/projected/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-kube-api-access-nkr4g\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.260927 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-config-data\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.266966 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-db-sync-config-data\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.268689 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-config-data\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.269326 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-combined-ca-bundle\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.279782 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkr4g\" (UniqueName: \"kubernetes.io/projected/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-kube-api-access-nkr4g\") pod \"glance-db-sync-rc66k\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.379453 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rc66k" Nov 27 11:55:41 crc kubenswrapper[4796]: I1127 11:55:41.948489 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rc66k"] Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.036991 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.176178 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-combined-ca-bundle\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.176346 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-scripts\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.176420 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/085a80b3-9f90-480a-8e26-fda7fb500c64-etc-swift\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.177550 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085a80b3-9f90-480a-8e26-fda7fb500c64-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.176505 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7mnl\" (UniqueName: \"kubernetes.io/projected/085a80b3-9f90-480a-8e26-fda7fb500c64-kube-api-access-g7mnl\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.177717 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-dispersionconf\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.177745 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-ring-data-devices\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.178547 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-swiftconf\") pod \"085a80b3-9f90-480a-8e26-fda7fb500c64\" (UID: \"085a80b3-9f90-480a-8e26-fda7fb500c64\") " Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.178142 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.179184 4796 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/085a80b3-9f90-480a-8e26-fda7fb500c64-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.179209 4796 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.180830 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085a80b3-9f90-480a-8e26-fda7fb500c64-kube-api-access-g7mnl" (OuterVolumeSpecName: "kube-api-access-g7mnl") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "kube-api-access-g7mnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.182908 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.198737 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.199810 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.203784 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-scripts" (OuterVolumeSpecName: "scripts") pod "085a80b3-9f90-480a-8e26-fda7fb500c64" (UID: "085a80b3-9f90-480a-8e26-fda7fb500c64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.282965 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.282989 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/085a80b3-9f90-480a-8e26-fda7fb500c64-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.282997 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7mnl\" (UniqueName: \"kubernetes.io/projected/085a80b3-9f90-480a-8e26-fda7fb500c64-kube-api-access-g7mnl\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.283016 4796 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.283026 4796 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/085a80b3-9f90-480a-8e26-fda7fb500c64-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.746865 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jw9v8" event={"ID":"085a80b3-9f90-480a-8e26-fda7fb500c64","Type":"ContainerDied","Data":"ce47ce5ae279ea3e18ef8644c8d62a361e5d3175fd66a57f5051a11b68efbc7e"} Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.747229 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce47ce5ae279ea3e18ef8644c8d62a361e5d3175fd66a57f5051a11b68efbc7e" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.747359 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jw9v8" Nov 27 11:55:42 crc kubenswrapper[4796]: I1127 11:55:42.756384 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rc66k" event={"ID":"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21","Type":"ContainerStarted","Data":"3f8e50323aee6cb522be101ed17519ff7d3b253aef27a5c17571a3f0cc916f11"} Nov 27 11:55:46 crc kubenswrapper[4796]: I1127 11:55:46.171916 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zjmwj" podUID="99ef6c07-3c37-4a97-8fbc-bd09a91617a0" containerName="ovn-controller" probeResult="failure" output=< Nov 27 11:55:46 crc kubenswrapper[4796]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 11:55:46 crc kubenswrapper[4796]: > Nov 27 11:55:46 crc kubenswrapper[4796]: I1127 11:55:46.248310 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:55:48 crc kubenswrapper[4796]: I1127 11:55:48.807988 4796 generic.go:334] "Generic (PLEG): container finished" podID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerID="c189a5bb7656fc2b2a2ccbdbd7bc017c1a6d0494b20cf5077e9aa8b7abe84c46" exitCode=0 Nov 27 11:55:48 crc kubenswrapper[4796]: I1127 11:55:48.808171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b853d8bd-9978-45a7-ad7c-b3217d4a93b3","Type":"ContainerDied","Data":"c189a5bb7656fc2b2a2ccbdbd7bc017c1a6d0494b20cf5077e9aa8b7abe84c46"} Nov 27 11:55:48 crc kubenswrapper[4796]: I1127 11:55:48.810237 4796 generic.go:334] "Generic (PLEG): container finished" podID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerID="5d683f5dd13d49637f10b186e2b1c13b5c81e507a7e715a4253a8484afc9e8c6" exitCode=0 Nov 27 11:55:48 crc kubenswrapper[4796]: I1127 11:55:48.810307 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e9cd74f-d96d-4be4-b448-01d027b34689","Type":"ContainerDied","Data":"5d683f5dd13d49637f10b186e2b1c13b5c81e507a7e715a4253a8484afc9e8c6"} Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.252694 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zjmwj" podUID="99ef6c07-3c37-4a97-8fbc-bd09a91617a0" containerName="ovn-controller" probeResult="failure" output=< Nov 27 11:55:51 crc kubenswrapper[4796]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 11:55:51 crc kubenswrapper[4796]: > Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.340627 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7cdqc" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.558040 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zjmwj-config-bplb2"] Nov 27 11:55:51 crc kubenswrapper[4796]: E1127 11:55:51.558555 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085a80b3-9f90-480a-8e26-fda7fb500c64" containerName="swift-ring-rebalance" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.558587 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="085a80b3-9f90-480a-8e26-fda7fb500c64" containerName="swift-ring-rebalance" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.558929 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="085a80b3-9f90-480a-8e26-fda7fb500c64" containerName="swift-ring-rebalance" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.559849 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.563764 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.569238 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:55:51 crc kubenswrapper[4796]: E1127 11:55:51.569518 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.581943 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zjmwj-config-bplb2"] Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.689593 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlvfn\" (UniqueName: \"kubernetes.io/projected/8999ca7d-64a1-4bcc-8621-73c6fff323dd-kube-api-access-nlvfn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.689732 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-additional-scripts\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.689764 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run-ovn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.689791 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.689823 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-scripts\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.689887 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-log-ovn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.791158 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-additional-scripts\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.791204 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run-ovn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.791226 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.791247 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-scripts\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.791298 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-log-ovn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.791344 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlvfn\" (UniqueName: \"kubernetes.io/projected/8999ca7d-64a1-4bcc-8621-73c6fff323dd-kube-api-access-nlvfn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.792436 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-additional-scripts\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.792710 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run-ovn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.792770 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.794344 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-log-ovn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.795127 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-scripts\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.811854 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlvfn\" (UniqueName: \"kubernetes.io/projected/8999ca7d-64a1-4bcc-8621-73c6fff323dd-kube-api-access-nlvfn\") pod \"ovn-controller-zjmwj-config-bplb2\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:51 crc kubenswrapper[4796]: I1127 11:55:51.895649 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.123440 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zjmwj-config-bplb2"] Nov 27 11:55:54 crc kubenswrapper[4796]: W1127 11:55:54.127217 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8999ca7d_64a1_4bcc_8621_73c6fff323dd.slice/crio-c1402880be103aeb86c57ffaad5ae4d940e5085f29a70023200aadd968ed4543 WatchSource:0}: Error finding container c1402880be103aeb86c57ffaad5ae4d940e5085f29a70023200aadd968ed4543: Status 404 returned error can't find the container with id c1402880be103aeb86c57ffaad5ae4d940e5085f29a70023200aadd968ed4543 Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.864999 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b853d8bd-9978-45a7-ad7c-b3217d4a93b3","Type":"ContainerStarted","Data":"0cd3cd3cb58ddab12d8a55ce4bcef4301b0141513e2fc0b859404aed61f14527"} Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.865532 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.867342 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e9cd74f-d96d-4be4-b448-01d027b34689","Type":"ContainerStarted","Data":"f89b18a271e227ede16b294522eb260aa74c47f021952b0da1f556debaf2c649"} Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.867563 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.869131 4796 generic.go:334] "Generic (PLEG): container finished" podID="8999ca7d-64a1-4bcc-8621-73c6fff323dd" containerID="64784d397077ad3eb24b4d2ade0b7470d38a9a88edf30e8be30366a9c0a4e6d4" exitCode=0 Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.869201 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj-config-bplb2" event={"ID":"8999ca7d-64a1-4bcc-8621-73c6fff323dd","Type":"ContainerDied","Data":"64784d397077ad3eb24b4d2ade0b7470d38a9a88edf30e8be30366a9c0a4e6d4"} Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.869221 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj-config-bplb2" event={"ID":"8999ca7d-64a1-4bcc-8621-73c6fff323dd","Type":"ContainerStarted","Data":"c1402880be103aeb86c57ffaad5ae4d940e5085f29a70023200aadd968ed4543"} Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.871007 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rc66k" event={"ID":"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21","Type":"ContainerStarted","Data":"05f28f23c9d44321b0d441bb0983ccfa294b974fe75a951c9b4218acf1018a51"} Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.894398 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=68.203868948 podStartE2EDuration="1m19.894376075s" podCreationTimestamp="2025-11-27 11:54:35 +0000 UTC" firstStartedPulling="2025-11-27 11:54:59.846555191 +0000 UTC m=+1817.364874109" lastFinishedPulling="2025-11-27 11:55:11.537062318 +0000 UTC m=+1829.055381236" observedRunningTime="2025-11-27 11:55:54.891369865 +0000 UTC m=+1872.409688783" watchObservedRunningTime="2025-11-27 11:55:54.894376075 +0000 UTC m=+1872.412695023" Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.933753 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=66.871520252 podStartE2EDuration="1m18.933728975s" podCreationTimestamp="2025-11-27 11:54:36 +0000 UTC" firstStartedPulling="2025-11-27 11:55:00.752067197 +0000 UTC m=+1818.270386125" lastFinishedPulling="2025-11-27 11:55:12.81427593 +0000 UTC m=+1830.332594848" observedRunningTime="2025-11-27 11:55:54.928610859 +0000 UTC m=+1872.446929777" watchObservedRunningTime="2025-11-27 11:55:54.933728975 +0000 UTC m=+1872.452047933" Nov 27 11:55:54 crc kubenswrapper[4796]: I1127 11:55:54.963498 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-rc66k" podStartSLOduration=2.097027349 podStartE2EDuration="13.963480619s" podCreationTimestamp="2025-11-27 11:55:41 +0000 UTC" firstStartedPulling="2025-11-27 11:55:41.962112953 +0000 UTC m=+1859.480431901" lastFinishedPulling="2025-11-27 11:55:53.828566253 +0000 UTC m=+1871.346885171" observedRunningTime="2025-11-27 11:55:54.957615423 +0000 UTC m=+1872.475934371" watchObservedRunningTime="2025-11-27 11:55:54.963480619 +0000 UTC m=+1872.481799527" Nov 27 11:55:55 crc kubenswrapper[4796]: I1127 11:55:55.771712 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:55 crc kubenswrapper[4796]: I1127 11:55:55.777745 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56f6dd56-8bbf-4610-b48c-898cda2e2aab-etc-swift\") pod \"swift-storage-0\" (UID: \"56f6dd56-8bbf-4610-b48c-898cda2e2aab\") " pod="openstack/swift-storage-0" Nov 27 11:55:55 crc kubenswrapper[4796]: I1127 11:55:55.831983 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.206722 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zjmwj" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.349091 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.482387 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlvfn\" (UniqueName: \"kubernetes.io/projected/8999ca7d-64a1-4bcc-8621-73c6fff323dd-kube-api-access-nlvfn\") pod \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.482845 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run-ovn\") pod \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.482939 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8999ca7d-64a1-4bcc-8621-73c6fff323dd" (UID: "8999ca7d-64a1-4bcc-8621-73c6fff323dd"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.482962 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-scripts\") pod \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483049 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run\") pod \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483157 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run" (OuterVolumeSpecName: "var-run") pod "8999ca7d-64a1-4bcc-8621-73c6fff323dd" (UID: "8999ca7d-64a1-4bcc-8621-73c6fff323dd"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483181 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-log-ovn\") pod \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483209 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-additional-scripts\") pod \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\" (UID: \"8999ca7d-64a1-4bcc-8621-73c6fff323dd\") " Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483255 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8999ca7d-64a1-4bcc-8621-73c6fff323dd" (UID: "8999ca7d-64a1-4bcc-8621-73c6fff323dd"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483547 4796 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483567 4796 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483577 4796 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8999ca7d-64a1-4bcc-8621-73c6fff323dd-var-run\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483852 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8999ca7d-64a1-4bcc-8621-73c6fff323dd" (UID: "8999ca7d-64a1-4bcc-8621-73c6fff323dd"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.483991 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-scripts" (OuterVolumeSpecName: "scripts") pod "8999ca7d-64a1-4bcc-8621-73c6fff323dd" (UID: "8999ca7d-64a1-4bcc-8621-73c6fff323dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.488583 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8999ca7d-64a1-4bcc-8621-73c6fff323dd-kube-api-access-nlvfn" (OuterVolumeSpecName: "kube-api-access-nlvfn") pod "8999ca7d-64a1-4bcc-8621-73c6fff323dd" (UID: "8999ca7d-64a1-4bcc-8621-73c6fff323dd"). InnerVolumeSpecName "kube-api-access-nlvfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.503937 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 27 11:55:56 crc kubenswrapper[4796]: W1127 11:55:56.505416 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56f6dd56_8bbf_4610_b48c_898cda2e2aab.slice/crio-fa689fc6142e89a557f4c257da56afb10789427787c2db91d5bc65b25f3204be WatchSource:0}: Error finding container fa689fc6142e89a557f4c257da56afb10789427787c2db91d5bc65b25f3204be: Status 404 returned error can't find the container with id fa689fc6142e89a557f4c257da56afb10789427787c2db91d5bc65b25f3204be Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.585185 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.585409 4796 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8999ca7d-64a1-4bcc-8621-73c6fff323dd-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.585530 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlvfn\" (UniqueName: \"kubernetes.io/projected/8999ca7d-64a1-4bcc-8621-73c6fff323dd-kube-api-access-nlvfn\") on node \"crc\" DevicePath \"\"" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.893520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"fa689fc6142e89a557f4c257da56afb10789427787c2db91d5bc65b25f3204be"} Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.894623 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj-config-bplb2" event={"ID":"8999ca7d-64a1-4bcc-8621-73c6fff323dd","Type":"ContainerDied","Data":"c1402880be103aeb86c57ffaad5ae4d940e5085f29a70023200aadd968ed4543"} Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.894663 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1402880be103aeb86c57ffaad5ae4d940e5085f29a70023200aadd968ed4543" Nov 27 11:55:56 crc kubenswrapper[4796]: I1127 11:55:56.894680 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-bplb2" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.470666 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zjmwj-config-bplb2"] Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.481811 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zjmwj-config-bplb2"] Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.578864 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8999ca7d-64a1-4bcc-8621-73c6fff323dd" path="/var/lib/kubelet/pods/8999ca7d-64a1-4bcc-8621-73c6fff323dd/volumes" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.585509 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zjmwj-config-4hlcw"] Nov 27 11:55:57 crc kubenswrapper[4796]: E1127 11:55:57.585855 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8999ca7d-64a1-4bcc-8621-73c6fff323dd" containerName="ovn-config" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.585872 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8999ca7d-64a1-4bcc-8621-73c6fff323dd" containerName="ovn-config" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.586024 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8999ca7d-64a1-4bcc-8621-73c6fff323dd" containerName="ovn-config" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.586565 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.591744 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.596209 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zjmwj-config-4hlcw"] Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.704096 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-scripts\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.704447 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run-ovn\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.704484 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-log-ovn\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.704529 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-additional-scripts\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.704549 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.704672 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m22wc\" (UniqueName: \"kubernetes.io/projected/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-kube-api-access-m22wc\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806389 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-log-ovn\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806482 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-additional-scripts\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806511 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806590 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m22wc\" (UniqueName: \"kubernetes.io/projected/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-kube-api-access-m22wc\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806617 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-scripts\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806658 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run-ovn\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806741 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-log-ovn\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806765 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run-ovn\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.806813 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.807355 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-additional-scripts\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.808622 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-scripts\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.830045 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m22wc\" (UniqueName: \"kubernetes.io/projected/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-kube-api-access-m22wc\") pod \"ovn-controller-zjmwj-config-4hlcw\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:57 crc kubenswrapper[4796]: I1127 11:55:57.956212 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:55:58 crc kubenswrapper[4796]: I1127 11:55:58.415966 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zjmwj-config-4hlcw"] Nov 27 11:55:58 crc kubenswrapper[4796]: I1127 11:55:58.924740 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj-config-4hlcw" event={"ID":"be09591e-b9a3-42ca-97fe-8f0b6a4640b8","Type":"ContainerStarted","Data":"e6a18f3385eb2bc3b367d9dc33f41ae00bc80a58e14c9aa7c7f4f431665508ff"} Nov 27 11:56:00 crc kubenswrapper[4796]: I1127 11:56:00.968011 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"40e1dcd1f93b9707dec282aae1e4bae2fd0753a4b748f80f54f542e40ac4d748"} Nov 27 11:56:00 crc kubenswrapper[4796]: I1127 11:56:00.972142 4796 generic.go:334] "Generic (PLEG): container finished" podID="be09591e-b9a3-42ca-97fe-8f0b6a4640b8" containerID="f1c6b477e6fdb687fadb6f5f399f44798a3996c714ef04a6abcd05ec70a55c89" exitCode=0 Nov 27 11:56:00 crc kubenswrapper[4796]: I1127 11:56:00.972196 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj-config-4hlcw" event={"ID":"be09591e-b9a3-42ca-97fe-8f0b6a4640b8","Type":"ContainerDied","Data":"f1c6b477e6fdb687fadb6f5f399f44798a3996c714ef04a6abcd05ec70a55c89"} Nov 27 11:56:01 crc kubenswrapper[4796]: I1127 11:56:01.984560 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"7a008e8f9e6f6855e473ef70a4820db0e0eeaf03bb1501ffad6cf43a2328008b"} Nov 27 11:56:01 crc kubenswrapper[4796]: I1127 11:56:01.986581 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"32e29955ba788445ea2a1e5e3cf9b15187571021647fde049f0489a15f01e8c0"} Nov 27 11:56:01 crc kubenswrapper[4796]: I1127 11:56:01.986790 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"daaaa639876fad998fcfb5766d0259fe674579d81d3d83b71453635c2526093c"} Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.342977 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401123 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run\") pod \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401207 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-log-ovn\") pod \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401244 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-scripts\") pod \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401342 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run-ovn\") pod \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401360 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-additional-scripts\") pod \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401482 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m22wc\" (UniqueName: \"kubernetes.io/projected/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-kube-api-access-m22wc\") pod \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\" (UID: \"be09591e-b9a3-42ca-97fe-8f0b6a4640b8\") " Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401505 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "be09591e-b9a3-42ca-97fe-8f0b6a4640b8" (UID: "be09591e-b9a3-42ca-97fe-8f0b6a4640b8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401583 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run" (OuterVolumeSpecName: "var-run") pod "be09591e-b9a3-42ca-97fe-8f0b6a4640b8" (UID: "be09591e-b9a3-42ca-97fe-8f0b6a4640b8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.401612 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "be09591e-b9a3-42ca-97fe-8f0b6a4640b8" (UID: "be09591e-b9a3-42ca-97fe-8f0b6a4640b8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.402177 4796 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.402207 4796 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.402220 4796 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.402669 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "be09591e-b9a3-42ca-97fe-8f0b6a4640b8" (UID: "be09591e-b9a3-42ca-97fe-8f0b6a4640b8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.403067 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-scripts" (OuterVolumeSpecName: "scripts") pod "be09591e-b9a3-42ca-97fe-8f0b6a4640b8" (UID: "be09591e-b9a3-42ca-97fe-8f0b6a4640b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.407462 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-kube-api-access-m22wc" (OuterVolumeSpecName: "kube-api-access-m22wc") pod "be09591e-b9a3-42ca-97fe-8f0b6a4640b8" (UID: "be09591e-b9a3-42ca-97fe-8f0b6a4640b8"). InnerVolumeSpecName "kube-api-access-m22wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.503805 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m22wc\" (UniqueName: \"kubernetes.io/projected/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-kube-api-access-m22wc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.503843 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:02 crc kubenswrapper[4796]: I1127 11:56:02.503854 4796 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/be09591e-b9a3-42ca-97fe-8f0b6a4640b8-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.002673 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zjmwj-config-4hlcw" event={"ID":"be09591e-b9a3-42ca-97fe-8f0b6a4640b8","Type":"ContainerDied","Data":"e6a18f3385eb2bc3b367d9dc33f41ae00bc80a58e14c9aa7c7f4f431665508ff"} Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.002721 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6a18f3385eb2bc3b367d9dc33f41ae00bc80a58e14c9aa7c7f4f431665508ff" Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.002796 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zjmwj-config-4hlcw" Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.010034 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"e664267787a0947056d61ec04a73c3c1dc6924de90f9ecd914643a0504586733"} Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.435165 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zjmwj-config-4hlcw"] Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.444558 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zjmwj-config-4hlcw"] Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.582776 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:56:03 crc kubenswrapper[4796]: E1127 11:56:03.583159 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:56:03 crc kubenswrapper[4796]: I1127 11:56:03.584222 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be09591e-b9a3-42ca-97fe-8f0b6a4640b8" path="/var/lib/kubelet/pods/be09591e-b9a3-42ca-97fe-8f0b6a4640b8/volumes" Nov 27 11:56:04 crc kubenswrapper[4796]: I1127 11:56:04.023192 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"b16d944a8a381dd1239806334081a14c616869c928edd706c030b9b7698292fa"} Nov 27 11:56:04 crc kubenswrapper[4796]: I1127 11:56:04.023238 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"4f97b0800feffc9ccd29ffac07a127bb6d89aec5db01e1e67d50e7a46d2c4b45"} Nov 27 11:56:04 crc kubenswrapper[4796]: I1127 11:56:04.023252 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"8e1ff8058582cb24b394bf4e38106641a6353f8b0f3e01ee87e390608f033ea4"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.060882 4796 generic.go:334] "Generic (PLEG): container finished" podID="79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" containerID="05f28f23c9d44321b0d441bb0983ccfa294b974fe75a951c9b4218acf1018a51" exitCode=0 Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.061008 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rc66k" event={"ID":"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21","Type":"ContainerDied","Data":"05f28f23c9d44321b0d441bb0983ccfa294b974fe75a951c9b4218acf1018a51"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.070028 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"b363f945f498ac809ef30853488f18fa46409ab08542b0cdfcb85467a4624d22"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.070057 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"64ef6e4b9f36f74086802bdb289515969f2e1c5a2755971dce5077bf600d6cc5"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.070066 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"6ff869e202b3225b1e75d25239bdf25a55220997dc72907db175ca28d4b9e728"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.070075 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"3fe560b62022f7fbde34e072c609d585c0296858194cfccf2669d9cc147a5302"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.070085 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"4f809065e7785a792a4ffc59f36adf172a102d9238302e1377caf1988d7143c0"} Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.193409 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.479501 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xr4s2"] Nov 27 11:56:07 crc kubenswrapper[4796]: E1127 11:56:07.479819 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be09591e-b9a3-42ca-97fe-8f0b6a4640b8" containerName="ovn-config" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.479834 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="be09591e-b9a3-42ca-97fe-8f0b6a4640b8" containerName="ovn-config" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.479994 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="be09591e-b9a3-42ca-97fe-8f0b6a4640b8" containerName="ovn-config" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.480481 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.504585 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xr4s2"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.595557 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sddq8\" (UniqueName: \"kubernetes.io/projected/430e7010-722c-4c22-9097-b65ed2975115-kube-api-access-sddq8\") pod \"cinder-db-create-xr4s2\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.595848 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/430e7010-722c-4c22-9097-b65ed2975115-operator-scripts\") pod \"cinder-db-create-xr4s2\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.623961 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-77aa-account-create-update-qqwnn"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.625117 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.627090 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.640260 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-77aa-account-create-update-qqwnn"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.700795 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv5zn\" (UniqueName: \"kubernetes.io/projected/e84a3d5c-420f-437c-9ba9-901f550d54a7-kube-api-access-rv5zn\") pod \"cinder-77aa-account-create-update-qqwnn\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.701128 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sddq8\" (UniqueName: \"kubernetes.io/projected/430e7010-722c-4c22-9097-b65ed2975115-kube-api-access-sddq8\") pod \"cinder-db-create-xr4s2\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.701233 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e84a3d5c-420f-437c-9ba9-901f550d54a7-operator-scripts\") pod \"cinder-77aa-account-create-update-qqwnn\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.701354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/430e7010-722c-4c22-9097-b65ed2975115-operator-scripts\") pod \"cinder-db-create-xr4s2\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.703194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/430e7010-722c-4c22-9097-b65ed2975115-operator-scripts\") pod \"cinder-db-create-xr4s2\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.724330 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-m7gr7"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.725702 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.727692 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m7gr7"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.736555 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a1e4-account-create-update-5hjzz"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.737748 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.745411 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.753066 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sddq8\" (UniqueName: \"kubernetes.io/projected/430e7010-722c-4c22-9097-b65ed2975115-kube-api-access-sddq8\") pod \"cinder-db-create-xr4s2\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.759440 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.772203 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a1e4-account-create-update-5hjzz"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.800896 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.802536 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kkfg\" (UniqueName: \"kubernetes.io/projected/e8ac8842-ecad-4b5d-a7e9-6daa85540618-kube-api-access-8kkfg\") pod \"barbican-a1e4-account-create-update-5hjzz\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.802603 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r7hg\" (UniqueName: \"kubernetes.io/projected/9938c442-847a-48f5-bcd4-7454e8a7c7e1-kube-api-access-8r7hg\") pod \"barbican-db-create-m7gr7\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.802669 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9938c442-847a-48f5-bcd4-7454e8a7c7e1-operator-scripts\") pod \"barbican-db-create-m7gr7\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.802702 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv5zn\" (UniqueName: \"kubernetes.io/projected/e84a3d5c-420f-437c-9ba9-901f550d54a7-kube-api-access-rv5zn\") pod \"cinder-77aa-account-create-update-qqwnn\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.802752 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ac8842-ecad-4b5d-a7e9-6daa85540618-operator-scripts\") pod \"barbican-a1e4-account-create-update-5hjzz\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.802814 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e84a3d5c-420f-437c-9ba9-901f550d54a7-operator-scripts\") pod \"cinder-77aa-account-create-update-qqwnn\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.805167 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e84a3d5c-420f-437c-9ba9-901f550d54a7-operator-scripts\") pod \"cinder-77aa-account-create-update-qqwnn\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.823379 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv5zn\" (UniqueName: \"kubernetes.io/projected/e84a3d5c-420f-437c-9ba9-901f550d54a7-kube-api-access-rv5zn\") pod \"cinder-77aa-account-create-update-qqwnn\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.904027 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kkfg\" (UniqueName: \"kubernetes.io/projected/e8ac8842-ecad-4b5d-a7e9-6daa85540618-kube-api-access-8kkfg\") pod \"barbican-a1e4-account-create-update-5hjzz\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.904093 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r7hg\" (UniqueName: \"kubernetes.io/projected/9938c442-847a-48f5-bcd4-7454e8a7c7e1-kube-api-access-8r7hg\") pod \"barbican-db-create-m7gr7\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.904134 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9938c442-847a-48f5-bcd4-7454e8a7c7e1-operator-scripts\") pod \"barbican-db-create-m7gr7\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.904185 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ac8842-ecad-4b5d-a7e9-6daa85540618-operator-scripts\") pod \"barbican-a1e4-account-create-update-5hjzz\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.905117 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ac8842-ecad-4b5d-a7e9-6daa85540618-operator-scripts\") pod \"barbican-a1e4-account-create-update-5hjzz\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.905993 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-25q46"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.906996 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-25q46" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.906002 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9938c442-847a-48f5-bcd4-7454e8a7c7e1-operator-scripts\") pod \"barbican-db-create-m7gr7\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.924112 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-25q46"] Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.931225 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r7hg\" (UniqueName: \"kubernetes.io/projected/9938c442-847a-48f5-bcd4-7454e8a7c7e1-kube-api-access-8r7hg\") pod \"barbican-db-create-m7gr7\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.932436 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kkfg\" (UniqueName: \"kubernetes.io/projected/e8ac8842-ecad-4b5d-a7e9-6daa85540618-kube-api-access-8kkfg\") pod \"barbican-a1e4-account-create-update-5hjzz\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.945358 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:07 crc kubenswrapper[4796]: I1127 11:56:07.996045 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-qgs76"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.007157 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.010448 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/590cb807-ad51-4176-a32c-0366c5d976c5-operator-scripts\") pod \"heat-db-create-25q46\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.013898 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m554q\" (UniqueName: \"kubernetes.io/projected/590cb807-ad51-4176-a32c-0366c5d976c5-kube-api-access-m554q\") pod \"heat-db-create-25q46\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.014645 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.016397 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.016635 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.018137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qgs76"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.020324 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rc9hr" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.046047 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.118349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-combined-ca-bundle\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.118423 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/590cb807-ad51-4176-a32c-0366c5d976c5-operator-scripts\") pod \"heat-db-create-25q46\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.118452 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m554q\" (UniqueName: \"kubernetes.io/projected/590cb807-ad51-4176-a32c-0366c5d976c5-kube-api-access-m554q\") pod \"heat-db-create-25q46\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.118487 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z67bq\" (UniqueName: \"kubernetes.io/projected/9833ccd1-3ec0-487a-9ba3-b536a7f25829-kube-api-access-z67bq\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.118554 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-config-data\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.119499 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/590cb807-ad51-4176-a32c-0366c5d976c5-operator-scripts\") pod \"heat-db-create-25q46\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.121184 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f34e-account-create-update-w5g78"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.121627 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.122989 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.128164 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.133779 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f34e-account-create-update-w5g78"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.161744 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m554q\" (UniqueName: \"kubernetes.io/projected/590cb807-ad51-4176-a32c-0366c5d976c5-kube-api-access-m554q\") pod \"heat-db-create-25q46\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.171082 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"919924c707d99c42d1bcca19545840c5ec0521b2cf0d7e1ba8505e0d96929dca"} Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.171321 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"56f6dd56-8bbf-4610-b48c-898cda2e2aab","Type":"ContainerStarted","Data":"4461c7d711691723411e9b5aecbc927bb17736b96e857f2656c9875098137bde"} Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.199102 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-23af-account-create-update-4vg5g"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.209118 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.211191 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.220668 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9g5n\" (UniqueName: \"kubernetes.io/projected/85e62fd9-c328-49cf-8f64-a61fb678a313-kube-api-access-p9g5n\") pod \"neutron-f34e-account-create-update-w5g78\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.220909 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-config-data\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.221501 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85e62fd9-c328-49cf-8f64-a61fb678a313-operator-scripts\") pod \"neutron-f34e-account-create-update-w5g78\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.221575 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-combined-ca-bundle\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.221704 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z67bq\" (UniqueName: \"kubernetes.io/projected/9833ccd1-3ec0-487a-9ba3-b536a7f25829-kube-api-access-z67bq\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.229382 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-config-data\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.231189 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-25q46" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.231285 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-combined-ca-bundle\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.232373 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-622c8"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.247242 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.249756 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z67bq\" (UniqueName: \"kubernetes.io/projected/9833ccd1-3ec0-487a-9ba3-b536a7f25829-kube-api-access-z67bq\") pod \"keystone-db-sync-qgs76\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.252467 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-622c8"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.264067 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-23af-account-create-update-4vg5g"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.267415 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.867443874 podStartE2EDuration="46.267396435s" podCreationTimestamp="2025-11-27 11:55:22 +0000 UTC" firstStartedPulling="2025-11-27 11:55:56.507521909 +0000 UTC m=+1874.025840837" lastFinishedPulling="2025-11-27 11:56:05.90747447 +0000 UTC m=+1883.425793398" observedRunningTime="2025-11-27 11:56:08.231131788 +0000 UTC m=+1885.749450706" watchObservedRunningTime="2025-11-27 11:56:08.267396435 +0000 UTC m=+1885.785715353" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.324845 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9g5n\" (UniqueName: \"kubernetes.io/projected/85e62fd9-c328-49cf-8f64-a61fb678a313-kube-api-access-p9g5n\") pod \"neutron-f34e-account-create-update-w5g78\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.324955 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp58k\" (UniqueName: \"kubernetes.io/projected/fc65938f-ce7e-462d-bd24-b8265723c981-kube-api-access-kp58k\") pod \"heat-23af-account-create-update-4vg5g\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.324984 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85e62fd9-c328-49cf-8f64-a61fb678a313-operator-scripts\") pod \"neutron-f34e-account-create-update-w5g78\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.325054 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgfc7\" (UniqueName: \"kubernetes.io/projected/9c223cd8-b92c-470b-bf11-f0c41d352887-kube-api-access-vgfc7\") pod \"neutron-db-create-622c8\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.325077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc65938f-ce7e-462d-bd24-b8265723c981-operator-scripts\") pod \"heat-23af-account-create-update-4vg5g\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.325101 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c223cd8-b92c-470b-bf11-f0c41d352887-operator-scripts\") pod \"neutron-db-create-622c8\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.326661 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85e62fd9-c328-49cf-8f64-a61fb678a313-operator-scripts\") pod \"neutron-f34e-account-create-update-w5g78\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.366101 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9g5n\" (UniqueName: \"kubernetes.io/projected/85e62fd9-c328-49cf-8f64-a61fb678a313-kube-api-access-p9g5n\") pod \"neutron-f34e-account-create-update-w5g78\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.426172 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc65938f-ce7e-462d-bd24-b8265723c981-operator-scripts\") pod \"heat-23af-account-create-update-4vg5g\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.426221 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c223cd8-b92c-470b-bf11-f0c41d352887-operator-scripts\") pod \"neutron-db-create-622c8\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.426325 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp58k\" (UniqueName: \"kubernetes.io/projected/fc65938f-ce7e-462d-bd24-b8265723c981-kube-api-access-kp58k\") pod \"heat-23af-account-create-update-4vg5g\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.426391 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgfc7\" (UniqueName: \"kubernetes.io/projected/9c223cd8-b92c-470b-bf11-f0c41d352887-kube-api-access-vgfc7\") pod \"neutron-db-create-622c8\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.427219 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c223cd8-b92c-470b-bf11-f0c41d352887-operator-scripts\") pod \"neutron-db-create-622c8\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.427549 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc65938f-ce7e-462d-bd24-b8265723c981-operator-scripts\") pod \"heat-23af-account-create-update-4vg5g\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.451421 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp58k\" (UniqueName: \"kubernetes.io/projected/fc65938f-ce7e-462d-bd24-b8265723c981-kube-api-access-kp58k\") pod \"heat-23af-account-create-update-4vg5g\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.454358 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgfc7\" (UniqueName: \"kubernetes.io/projected/9c223cd8-b92c-470b-bf11-f0c41d352887-kube-api-access-vgfc7\") pod \"neutron-db-create-622c8\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.465743 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xr4s2"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.487760 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.510001 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.534427 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.565447 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-622c8" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.610235 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lsb5x"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.614580 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.619488 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.628954 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lsb5x"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.645078 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m7gr7"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.691356 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-77aa-account-create-update-qqwnn"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.809648 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-25q46"] Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.843043 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-config\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.843452 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.843518 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.843559 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.843585 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.843632 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6p8s\" (UniqueName: \"kubernetes.io/projected/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-kube-api-access-b6p8s\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.888093 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rc66k" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947305 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-config-data\") pod \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947344 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-db-sync-config-data\") pod \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947388 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkr4g\" (UniqueName: \"kubernetes.io/projected/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-kube-api-access-nkr4g\") pod \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947447 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-combined-ca-bundle\") pod \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\" (UID: \"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21\") " Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-config\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947773 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947830 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947869 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947895 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.947940 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6p8s\" (UniqueName: \"kubernetes.io/projected/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-kube-api-access-b6p8s\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.954991 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.956081 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.956194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.956586 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-config\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.961665 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.976706 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" (UID: "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.978458 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6p8s\" (UniqueName: \"kubernetes.io/projected/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-kube-api-access-b6p8s\") pod \"dnsmasq-dns-5c79d794d7-lsb5x\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:08 crc kubenswrapper[4796]: I1127 11:56:08.984083 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-kube-api-access-nkr4g" (OuterVolumeSpecName: "kube-api-access-nkr4g") pod "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" (UID: "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21"). InnerVolumeSpecName "kube-api-access-nkr4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.028622 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a1e4-account-create-update-5hjzz"] Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.033970 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" (UID: "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.049029 4796 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.049049 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkr4g\" (UniqueName: \"kubernetes.io/projected/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-kube-api-access-nkr4g\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.049059 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.099084 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-config-data" (OuterVolumeSpecName: "config-data") pod "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" (UID: "79e9fbbf-a929-4d4a-8fb9-ab083c0aea21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.152480 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.188901 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a1e4-account-create-update-5hjzz" event={"ID":"e8ac8842-ecad-4b5d-a7e9-6daa85540618","Type":"ContainerStarted","Data":"a0723462ce24c7edd0922e274c7a9271324594444f78a43ddfb50e128e89807d"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.194888 4796 generic.go:334] "Generic (PLEG): container finished" podID="430e7010-722c-4c22-9097-b65ed2975115" containerID="b578ec53d5f500a9fb7067ffdeed3cd4222a51aa161b02a1f7f705fa980b5a12" exitCode=0 Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.195016 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xr4s2" event={"ID":"430e7010-722c-4c22-9097-b65ed2975115","Type":"ContainerDied","Data":"b578ec53d5f500a9fb7067ffdeed3cd4222a51aa161b02a1f7f705fa980b5a12"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.195052 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xr4s2" event={"ID":"430e7010-722c-4c22-9097-b65ed2975115","Type":"ContainerStarted","Data":"334c3452a564399a05c6eb139ed60f977c7c71f5ef561c846a81d3585221139e"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.206466 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m7gr7" event={"ID":"9938c442-847a-48f5-bcd4-7454e8a7c7e1","Type":"ContainerStarted","Data":"7741b2af51aca03e3d3d35359f0ead06621672ac702eaadaa751bb52ed132e4d"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.206510 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m7gr7" event={"ID":"9938c442-847a-48f5-bcd4-7454e8a7c7e1","Type":"ContainerStarted","Data":"ff0f44346d76e547be76d5e68c0924de3a115f4e2a3710c625dd3c66a6801c91"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.207824 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-25q46" event={"ID":"590cb807-ad51-4176-a32c-0366c5d976c5","Type":"ContainerStarted","Data":"3efe961e11b78b681de5df06f0131550b232523beda6ae7f9295a4ebf3bb441d"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.217671 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-77aa-account-create-update-qqwnn" event={"ID":"e84a3d5c-420f-437c-9ba9-901f550d54a7","Type":"ContainerStarted","Data":"75892ccedad2f51d2021de4b3d7ab47f62dac7c2e87e1399591fc083cba8ea67"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.217718 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-77aa-account-create-update-qqwnn" event={"ID":"e84a3d5c-420f-437c-9ba9-901f550d54a7","Type":"ContainerStarted","Data":"ba5a27e2e99dd7786faba67703b4a4d671dc066e1384d707bbb8cb4272afcefc"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.233493 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-m7gr7" podStartSLOduration=2.233472697 podStartE2EDuration="2.233472697s" podCreationTimestamp="2025-11-27 11:56:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:09.22907465 +0000 UTC m=+1886.747393568" watchObservedRunningTime="2025-11-27 11:56:09.233472697 +0000 UTC m=+1886.751791615" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.237605 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rc66k" event={"ID":"79e9fbbf-a929-4d4a-8fb9-ab083c0aea21","Type":"ContainerDied","Data":"3f8e50323aee6cb522be101ed17519ff7d3b253aef27a5c17571a3f0cc916f11"} Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.237647 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f8e50323aee6cb522be101ed17519ff7d3b253aef27a5c17571a3f0cc916f11" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.237822 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rc66k" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.254750 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-77aa-account-create-update-qqwnn" podStartSLOduration=2.254728174 podStartE2EDuration="2.254728174s" podCreationTimestamp="2025-11-27 11:56:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:09.247041019 +0000 UTC m=+1886.765359937" watchObservedRunningTime="2025-11-27 11:56:09.254728174 +0000 UTC m=+1886.773047092" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.255058 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.298221 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qgs76"] Nov 27 11:56:09 crc kubenswrapper[4796]: W1127 11:56:09.300480 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9833ccd1_3ec0_487a_9ba3_b536a7f25829.slice/crio-970c0af0a3dbd1ba42203571e0beb9e6fec7cec1c13ce7ff833f61fa7a0a302e WatchSource:0}: Error finding container 970c0af0a3dbd1ba42203571e0beb9e6fec7cec1c13ce7ff833f61fa7a0a302e: Status 404 returned error can't find the container with id 970c0af0a3dbd1ba42203571e0beb9e6fec7cec1c13ce7ff833f61fa7a0a302e Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.323438 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-23af-account-create-update-4vg5g"] Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.404184 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-622c8"] Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.469369 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f34e-account-create-update-w5g78"] Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.494415 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lsb5x"] Nov 27 11:56:09 crc kubenswrapper[4796]: W1127 11:56:09.539299 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85e62fd9_c328_49cf_8f64_a61fb678a313.slice/crio-fd0d1208f77bd1711c0394c6103950e973ea5939e577975263a8b368bd7877be WatchSource:0}: Error finding container fd0d1208f77bd1711c0394c6103950e973ea5939e577975263a8b368bd7877be: Status 404 returned error can't find the container with id fd0d1208f77bd1711c0394c6103950e973ea5939e577975263a8b368bd7877be Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.554169 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ngtn7"] Nov 27 11:56:09 crc kubenswrapper[4796]: E1127 11:56:09.554546 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" containerName="glance-db-sync" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.554558 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" containerName="glance-db-sync" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.554792 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" containerName="glance-db-sync" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.555693 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.622894 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ngtn7"] Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.664800 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.664855 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.664886 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.664911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqjrk\" (UniqueName: \"kubernetes.io/projected/b5504fa9-43f7-442d-a025-245928896378-kube-api-access-dqjrk\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.664940 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.665008 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-config\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.766442 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.766491 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.766522 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqjrk\" (UniqueName: \"kubernetes.io/projected/b5504fa9-43f7-442d-a025-245928896378-kube-api-access-dqjrk\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.766554 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.766622 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-config\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.766662 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.768980 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.769468 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.769877 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-config\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.770377 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.771286 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:09 crc kubenswrapper[4796]: I1127 11:56:09.791527 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqjrk\" (UniqueName: \"kubernetes.io/projected/b5504fa9-43f7-442d-a025-245928896378-kube-api-access-dqjrk\") pod \"dnsmasq-dns-5f59b8f679-ngtn7\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.008741 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.090092 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lsb5x"] Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.248076 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" event={"ID":"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa","Type":"ContainerStarted","Data":"be1c7b7eabcd7718776a6ed8a0fa6c18312b9ae2337a9f82eb0afac64f7ed8de"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.251079 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a1e4-account-create-update-5hjzz" event={"ID":"e8ac8842-ecad-4b5d-a7e9-6daa85540618","Type":"ContainerStarted","Data":"364ff08d502464b849bd6befb4c6bded68e2c1e9ecd53e9a77f5d0f7b493c170"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.268734 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-a1e4-account-create-update-5hjzz" podStartSLOduration=3.268718874 podStartE2EDuration="3.268718874s" podCreationTimestamp="2025-11-27 11:56:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:10.266014112 +0000 UTC m=+1887.784333030" watchObservedRunningTime="2025-11-27 11:56:10.268718874 +0000 UTC m=+1887.787037792" Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.276059 4796 generic.go:334] "Generic (PLEG): container finished" podID="e84a3d5c-420f-437c-9ba9-901f550d54a7" containerID="75892ccedad2f51d2021de4b3d7ab47f62dac7c2e87e1399591fc083cba8ea67" exitCode=0 Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.276174 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-77aa-account-create-update-qqwnn" event={"ID":"e84a3d5c-420f-437c-9ba9-901f550d54a7","Type":"ContainerDied","Data":"75892ccedad2f51d2021de4b3d7ab47f62dac7c2e87e1399591fc083cba8ea67"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.284976 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-622c8" event={"ID":"9c223cd8-b92c-470b-bf11-f0c41d352887","Type":"ContainerStarted","Data":"7774f31c04e4de76576e1f215c57f1bca6c0ac373466f09fdd67a64c9d2f0b45"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.285010 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-622c8" event={"ID":"9c223cd8-b92c-470b-bf11-f0c41d352887","Type":"ContainerStarted","Data":"42637e1c776ecdada5919c29af5fe759457dcd801110ea1b656304ee64cf9cae"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.287393 4796 generic.go:334] "Generic (PLEG): container finished" podID="590cb807-ad51-4176-a32c-0366c5d976c5" containerID="deb31ef07e4a5dfc949da1fc2b761a3a9035fcbff0cf81587d0b972e4e1681a5" exitCode=0 Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.287473 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-25q46" event={"ID":"590cb807-ad51-4176-a32c-0366c5d976c5","Type":"ContainerDied","Data":"deb31ef07e4a5dfc949da1fc2b761a3a9035fcbff0cf81587d0b972e4e1681a5"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.289174 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-23af-account-create-update-4vg5g" event={"ID":"fc65938f-ce7e-462d-bd24-b8265723c981","Type":"ContainerStarted","Data":"4c8f5b43130fd3a0ee0a2575e5c9f40c418136a30c16b319e0eeff5e7f5043ac"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.289207 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-23af-account-create-update-4vg5g" event={"ID":"fc65938f-ce7e-462d-bd24-b8265723c981","Type":"ContainerStarted","Data":"a0f42e263521e642462b3cc44531a41d6c9d14905948bcfe626a339cc5bc63ab"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.293825 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qgs76" event={"ID":"9833ccd1-3ec0-487a-9ba3-b536a7f25829","Type":"ContainerStarted","Data":"970c0af0a3dbd1ba42203571e0beb9e6fec7cec1c13ce7ff833f61fa7a0a302e"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.299638 4796 generic.go:334] "Generic (PLEG): container finished" podID="9938c442-847a-48f5-bcd4-7454e8a7c7e1" containerID="7741b2af51aca03e3d3d35359f0ead06621672ac702eaadaa751bb52ed132e4d" exitCode=0 Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.299746 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m7gr7" event={"ID":"9938c442-847a-48f5-bcd4-7454e8a7c7e1","Type":"ContainerDied","Data":"7741b2af51aca03e3d3d35359f0ead06621672ac702eaadaa751bb52ed132e4d"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.301953 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f34e-account-create-update-w5g78" event={"ID":"85e62fd9-c328-49cf-8f64-a61fb678a313","Type":"ContainerStarted","Data":"93aab9689a1f5e7f1df961b11a2fad20b297a011daa2152155e216b659ea760d"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.301982 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f34e-account-create-update-w5g78" event={"ID":"85e62fd9-c328-49cf-8f64-a61fb678a313","Type":"ContainerStarted","Data":"fd0d1208f77bd1711c0394c6103950e973ea5939e577975263a8b368bd7877be"} Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.307422 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-622c8" podStartSLOduration=2.307407756 podStartE2EDuration="2.307407756s" podCreationTimestamp="2025-11-27 11:56:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:10.304711154 +0000 UTC m=+1887.823030072" watchObservedRunningTime="2025-11-27 11:56:10.307407756 +0000 UTC m=+1887.825726674" Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.350631 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-23af-account-create-update-4vg5g" podStartSLOduration=2.350613829 podStartE2EDuration="2.350613829s" podCreationTimestamp="2025-11-27 11:56:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:10.344470705 +0000 UTC m=+1887.862789623" watchObservedRunningTime="2025-11-27 11:56:10.350613829 +0000 UTC m=+1887.868932737" Nov 27 11:56:10 crc kubenswrapper[4796]: I1127 11:56:10.371182 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f34e-account-create-update-w5g78" podStartSLOduration=2.371165397 podStartE2EDuration="2.371165397s" podCreationTimestamp="2025-11-27 11:56:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:10.363931964 +0000 UTC m=+1887.882250882" watchObservedRunningTime="2025-11-27 11:56:10.371165397 +0000 UTC m=+1887.889484315" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.574518 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ngtn7"] Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.601785 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.788462 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/430e7010-722c-4c22-9097-b65ed2975115-operator-scripts\") pod \"430e7010-722c-4c22-9097-b65ed2975115\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.788710 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sddq8\" (UniqueName: \"kubernetes.io/projected/430e7010-722c-4c22-9097-b65ed2975115-kube-api-access-sddq8\") pod \"430e7010-722c-4c22-9097-b65ed2975115\" (UID: \"430e7010-722c-4c22-9097-b65ed2975115\") " Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.789678 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430e7010-722c-4c22-9097-b65ed2975115-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "430e7010-722c-4c22-9097-b65ed2975115" (UID: "430e7010-722c-4c22-9097-b65ed2975115"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.792560 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/430e7010-722c-4c22-9097-b65ed2975115-kube-api-access-sddq8" (OuterVolumeSpecName: "kube-api-access-sddq8") pod "430e7010-722c-4c22-9097-b65ed2975115" (UID: "430e7010-722c-4c22-9097-b65ed2975115"). InnerVolumeSpecName "kube-api-access-sddq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.896155 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sddq8\" (UniqueName: \"kubernetes.io/projected/430e7010-722c-4c22-9097-b65ed2975115-kube-api-access-sddq8\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:10.896185 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/430e7010-722c-4c22-9097-b65ed2975115-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.311634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xr4s2" event={"ID":"430e7010-722c-4c22-9097-b65ed2975115","Type":"ContainerDied","Data":"334c3452a564399a05c6eb139ed60f977c7c71f5ef561c846a81d3585221139e"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.311945 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="334c3452a564399a05c6eb139ed60f977c7c71f5ef561c846a81d3585221139e" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.311859 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xr4s2" Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.313678 4796 generic.go:334] "Generic (PLEG): container finished" podID="85e62fd9-c328-49cf-8f64-a61fb678a313" containerID="93aab9689a1f5e7f1df961b11a2fad20b297a011daa2152155e216b659ea760d" exitCode=0 Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.313730 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f34e-account-create-update-w5g78" event={"ID":"85e62fd9-c328-49cf-8f64-a61fb678a313","Type":"ContainerDied","Data":"93aab9689a1f5e7f1df961b11a2fad20b297a011daa2152155e216b659ea760d"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.318524 4796 generic.go:334] "Generic (PLEG): container finished" podID="9c223cd8-b92c-470b-bf11-f0c41d352887" containerID="7774f31c04e4de76576e1f215c57f1bca6c0ac373466f09fdd67a64c9d2f0b45" exitCode=0 Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.318570 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-622c8" event={"ID":"9c223cd8-b92c-470b-bf11-f0c41d352887","Type":"ContainerDied","Data":"7774f31c04e4de76576e1f215c57f1bca6c0ac373466f09fdd67a64c9d2f0b45"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.320098 4796 generic.go:334] "Generic (PLEG): container finished" podID="fc65938f-ce7e-462d-bd24-b8265723c981" containerID="4c8f5b43130fd3a0ee0a2575e5c9f40c418136a30c16b319e0eeff5e7f5043ac" exitCode=0 Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.320136 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-23af-account-create-update-4vg5g" event={"ID":"fc65938f-ce7e-462d-bd24-b8265723c981","Type":"ContainerDied","Data":"4c8f5b43130fd3a0ee0a2575e5c9f40c418136a30c16b319e0eeff5e7f5043ac"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.321949 4796 generic.go:334] "Generic (PLEG): container finished" podID="b5504fa9-43f7-442d-a025-245928896378" containerID="3b64e2db5966a560814b7e9195f9f68f793526b34430de018558d8d04b2c193b" exitCode=0 Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.321983 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" event={"ID":"b5504fa9-43f7-442d-a025-245928896378","Type":"ContainerDied","Data":"3b64e2db5966a560814b7e9195f9f68f793526b34430de018558d8d04b2c193b"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.321998 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" event={"ID":"b5504fa9-43f7-442d-a025-245928896378","Type":"ContainerStarted","Data":"b43aa1c084be18beb0c7d8c6ab9e127070bc6157d32a78cce808f9aa3a62392d"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.329948 4796 generic.go:334] "Generic (PLEG): container finished" podID="a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" containerID="5298ab6262579a9e66660d394bd60a96b6c18979aee0cbfc2b7986cb330dc962" exitCode=0 Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.329995 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" event={"ID":"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa","Type":"ContainerDied","Data":"5298ab6262579a9e66660d394bd60a96b6c18979aee0cbfc2b7986cb330dc962"} Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.334186 4796 generic.go:334] "Generic (PLEG): container finished" podID="e8ac8842-ecad-4b5d-a7e9-6daa85540618" containerID="364ff08d502464b849bd6befb4c6bded68e2c1e9ecd53e9a77f5d0f7b493c170" exitCode=0 Nov 27 11:56:11 crc kubenswrapper[4796]: I1127 11:56:11.334299 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a1e4-account-create-update-5hjzz" event={"ID":"e8ac8842-ecad-4b5d-a7e9-6daa85540618","Type":"ContainerDied","Data":"364ff08d502464b849bd6befb4c6bded68e2c1e9ecd53e9a77f5d0f7b493c170"} Nov 27 11:56:12 crc kubenswrapper[4796]: I1127 11:56:12.345144 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" event={"ID":"b5504fa9-43f7-442d-a025-245928896378","Type":"ContainerStarted","Data":"34355bf4aac94f65b33f4030a233e523b4d7e7a9d1bb86f41f18d35a702e77bf"} Nov 27 11:56:12 crc kubenswrapper[4796]: I1127 11:56:12.345849 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:12 crc kubenswrapper[4796]: I1127 11:56:12.369097 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" podStartSLOduration=3.369080265 podStartE2EDuration="3.369080265s" podCreationTimestamp="2025-11-27 11:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:12.360776204 +0000 UTC m=+1889.879095122" watchObservedRunningTime="2025-11-27 11:56:12.369080265 +0000 UTC m=+1889.887399183" Nov 27 11:56:14 crc kubenswrapper[4796]: I1127 11:56:14.988446 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.076568 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r7hg\" (UniqueName: \"kubernetes.io/projected/9938c442-847a-48f5-bcd4-7454e8a7c7e1-kube-api-access-8r7hg\") pod \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.076742 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9938c442-847a-48f5-bcd4-7454e8a7c7e1-operator-scripts\") pod \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\" (UID: \"9938c442-847a-48f5-bcd4-7454e8a7c7e1\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.077385 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9938c442-847a-48f5-bcd4-7454e8a7c7e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9938c442-847a-48f5-bcd4-7454e8a7c7e1" (UID: "9938c442-847a-48f5-bcd4-7454e8a7c7e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.081091 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9938c442-847a-48f5-bcd4-7454e8a7c7e1-kube-api-access-8r7hg" (OuterVolumeSpecName: "kube-api-access-8r7hg") pod "9938c442-847a-48f5-bcd4-7454e8a7c7e1" (UID: "9938c442-847a-48f5-bcd4-7454e8a7c7e1"). InnerVolumeSpecName "kube-api-access-8r7hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.108797 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.178246 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ac8842-ecad-4b5d-a7e9-6daa85540618-operator-scripts\") pod \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.178739 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kkfg\" (UniqueName: \"kubernetes.io/projected/e8ac8842-ecad-4b5d-a7e9-6daa85540618-kube-api-access-8kkfg\") pod \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\" (UID: \"e8ac8842-ecad-4b5d-a7e9-6daa85540618\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.178872 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ac8842-ecad-4b5d-a7e9-6daa85540618-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8ac8842-ecad-4b5d-a7e9-6daa85540618" (UID: "e8ac8842-ecad-4b5d-a7e9-6daa85540618"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.179262 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r7hg\" (UniqueName: \"kubernetes.io/projected/9938c442-847a-48f5-bcd4-7454e8a7c7e1-kube-api-access-8r7hg\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.179308 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9938c442-847a-48f5-bcd4-7454e8a7c7e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.179319 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ac8842-ecad-4b5d-a7e9-6daa85540618-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.183680 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ac8842-ecad-4b5d-a7e9-6daa85540618-kube-api-access-8kkfg" (OuterVolumeSpecName: "kube-api-access-8kkfg") pod "e8ac8842-ecad-4b5d-a7e9-6daa85540618" (UID: "e8ac8842-ecad-4b5d-a7e9-6daa85540618"). InnerVolumeSpecName "kube-api-access-8kkfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.186044 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.222579 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-25q46" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.243730 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.251784 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.267524 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.274701 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-622c8" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280146 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-swift-storage-0\") pod \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280355 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp58k\" (UniqueName: \"kubernetes.io/projected/fc65938f-ce7e-462d-bd24-b8265723c981-kube-api-access-kp58k\") pod \"fc65938f-ce7e-462d-bd24-b8265723c981\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280393 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/590cb807-ad51-4176-a32c-0366c5d976c5-operator-scripts\") pod \"590cb807-ad51-4176-a32c-0366c5d976c5\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280417 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-sb\") pod \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280502 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-nb\") pod \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280527 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv5zn\" (UniqueName: \"kubernetes.io/projected/e84a3d5c-420f-437c-9ba9-901f550d54a7-kube-api-access-rv5zn\") pod \"e84a3d5c-420f-437c-9ba9-901f550d54a7\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280562 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9g5n\" (UniqueName: \"kubernetes.io/projected/85e62fd9-c328-49cf-8f64-a61fb678a313-kube-api-access-p9g5n\") pod \"85e62fd9-c328-49cf-8f64-a61fb678a313\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280590 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e84a3d5c-420f-437c-9ba9-901f550d54a7-operator-scripts\") pod \"e84a3d5c-420f-437c-9ba9-901f550d54a7\" (UID: \"e84a3d5c-420f-437c-9ba9-901f550d54a7\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280620 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6p8s\" (UniqueName: \"kubernetes.io/projected/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-kube-api-access-b6p8s\") pod \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280648 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-svc\") pod \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280724 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-config\") pod \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\" (UID: \"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280762 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m554q\" (UniqueName: \"kubernetes.io/projected/590cb807-ad51-4176-a32c-0366c5d976c5-kube-api-access-m554q\") pod \"590cb807-ad51-4176-a32c-0366c5d976c5\" (UID: \"590cb807-ad51-4176-a32c-0366c5d976c5\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280836 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85e62fd9-c328-49cf-8f64-a61fb678a313-operator-scripts\") pod \"85e62fd9-c328-49cf-8f64-a61fb678a313\" (UID: \"85e62fd9-c328-49cf-8f64-a61fb678a313\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.280886 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc65938f-ce7e-462d-bd24-b8265723c981-operator-scripts\") pod \"fc65938f-ce7e-462d-bd24-b8265723c981\" (UID: \"fc65938f-ce7e-462d-bd24-b8265723c981\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.281315 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e84a3d5c-420f-437c-9ba9-901f550d54a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e84a3d5c-420f-437c-9ba9-901f550d54a7" (UID: "e84a3d5c-420f-437c-9ba9-901f550d54a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.281395 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kkfg\" (UniqueName: \"kubernetes.io/projected/e8ac8842-ecad-4b5d-a7e9-6daa85540618-kube-api-access-8kkfg\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.281430 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85e62fd9-c328-49cf-8f64-a61fb678a313-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85e62fd9-c328-49cf-8f64-a61fb678a313" (UID: "85e62fd9-c328-49cf-8f64-a61fb678a313"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.281004 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/590cb807-ad51-4176-a32c-0366c5d976c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "590cb807-ad51-4176-a32c-0366c5d976c5" (UID: "590cb807-ad51-4176-a32c-0366c5d976c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.282091 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc65938f-ce7e-462d-bd24-b8265723c981-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc65938f-ce7e-462d-bd24-b8265723c981" (UID: "fc65938f-ce7e-462d-bd24-b8265723c981"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.285836 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc65938f-ce7e-462d-bd24-b8265723c981-kube-api-access-kp58k" (OuterVolumeSpecName: "kube-api-access-kp58k") pod "fc65938f-ce7e-462d-bd24-b8265723c981" (UID: "fc65938f-ce7e-462d-bd24-b8265723c981"). InnerVolumeSpecName "kube-api-access-kp58k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.308706 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" (UID: "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.323942 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" (UID: "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.327730 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-config" (OuterVolumeSpecName: "config") pod "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" (UID: "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.333753 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-kube-api-access-b6p8s" (OuterVolumeSpecName: "kube-api-access-b6p8s") pod "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" (UID: "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa"). InnerVolumeSpecName "kube-api-access-b6p8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.333809 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e84a3d5c-420f-437c-9ba9-901f550d54a7-kube-api-access-rv5zn" (OuterVolumeSpecName: "kube-api-access-rv5zn") pod "e84a3d5c-420f-437c-9ba9-901f550d54a7" (UID: "e84a3d5c-420f-437c-9ba9-901f550d54a7"). InnerVolumeSpecName "kube-api-access-rv5zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.333982 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/590cb807-ad51-4176-a32c-0366c5d976c5-kube-api-access-m554q" (OuterVolumeSpecName: "kube-api-access-m554q") pod "590cb807-ad51-4176-a32c-0366c5d976c5" (UID: "590cb807-ad51-4176-a32c-0366c5d976c5"). InnerVolumeSpecName "kube-api-access-m554q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.334437 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" (UID: "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.334467 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" (UID: "a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.338096 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e62fd9-c328-49cf-8f64-a61fb678a313-kube-api-access-p9g5n" (OuterVolumeSpecName: "kube-api-access-p9g5n") pod "85e62fd9-c328-49cf-8f64-a61fb678a313" (UID: "85e62fd9-c328-49cf-8f64-a61fb678a313"). InnerVolumeSpecName "kube-api-access-p9g5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.374108 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-77aa-account-create-update-qqwnn" event={"ID":"e84a3d5c-420f-437c-9ba9-901f550d54a7","Type":"ContainerDied","Data":"ba5a27e2e99dd7786faba67703b4a4d671dc066e1384d707bbb8cb4272afcefc"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.374158 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba5a27e2e99dd7786faba67703b4a4d671dc066e1384d707bbb8cb4272afcefc" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.374221 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-77aa-account-create-update-qqwnn" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.384999 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.385565 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgfc7\" (UniqueName: \"kubernetes.io/projected/9c223cd8-b92c-470b-bf11-f0c41d352887-kube-api-access-vgfc7\") pod \"9c223cd8-b92c-470b-bf11-f0c41d352887\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.385700 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c223cd8-b92c-470b-bf11-f0c41d352887-operator-scripts\") pod \"9c223cd8-b92c-470b-bf11-f0c41d352887\" (UID: \"9c223cd8-b92c-470b-bf11-f0c41d352887\") " Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386125 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp58k\" (UniqueName: \"kubernetes.io/projected/fc65938f-ce7e-462d-bd24-b8265723c981-kube-api-access-kp58k\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386148 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/590cb807-ad51-4176-a32c-0366c5d976c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386164 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386181 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386197 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv5zn\" (UniqueName: \"kubernetes.io/projected/e84a3d5c-420f-437c-9ba9-901f550d54a7-kube-api-access-rv5zn\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386213 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9g5n\" (UniqueName: \"kubernetes.io/projected/85e62fd9-c328-49cf-8f64-a61fb678a313-kube-api-access-p9g5n\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386229 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e84a3d5c-420f-437c-9ba9-901f550d54a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386242 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386254 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6p8s\" (UniqueName: \"kubernetes.io/projected/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-kube-api-access-b6p8s\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386285 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386296 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m554q\" (UniqueName: \"kubernetes.io/projected/590cb807-ad51-4176-a32c-0366c5d976c5-kube-api-access-m554q\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386310 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85e62fd9-c328-49cf-8f64-a61fb678a313-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386322 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc65938f-ce7e-462d-bd24-b8265723c981-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386334 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.386820 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c223cd8-b92c-470b-bf11-f0c41d352887-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c223cd8-b92c-470b-bf11-f0c41d352887" (UID: "9c223cd8-b92c-470b-bf11-f0c41d352887"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.387816 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lsb5x" event={"ID":"a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa","Type":"ContainerDied","Data":"be1c7b7eabcd7718776a6ed8a0fa6c18312b9ae2337a9f82eb0afac64f7ed8de"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.387901 4796 scope.go:117] "RemoveContainer" containerID="5298ab6262579a9e66660d394bd60a96b6c18979aee0cbfc2b7986cb330dc962" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.391086 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c223cd8-b92c-470b-bf11-f0c41d352887-kube-api-access-vgfc7" (OuterVolumeSpecName: "kube-api-access-vgfc7") pod "9c223cd8-b92c-470b-bf11-f0c41d352887" (UID: "9c223cd8-b92c-470b-bf11-f0c41d352887"). InnerVolumeSpecName "kube-api-access-vgfc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.400703 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a1e4-account-create-update-5hjzz" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.400716 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a1e4-account-create-update-5hjzz" event={"ID":"e8ac8842-ecad-4b5d-a7e9-6daa85540618","Type":"ContainerDied","Data":"a0723462ce24c7edd0922e274c7a9271324594444f78a43ddfb50e128e89807d"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.400756 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0723462ce24c7edd0922e274c7a9271324594444f78a43ddfb50e128e89807d" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.403613 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m7gr7" event={"ID":"9938c442-847a-48f5-bcd4-7454e8a7c7e1","Type":"ContainerDied","Data":"ff0f44346d76e547be76d5e68c0924de3a115f4e2a3710c625dd3c66a6801c91"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.403651 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff0f44346d76e547be76d5e68c0924de3a115f4e2a3710c625dd3c66a6801c91" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.403703 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m7gr7" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.405962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f34e-account-create-update-w5g78" event={"ID":"85e62fd9-c328-49cf-8f64-a61fb678a313","Type":"ContainerDied","Data":"fd0d1208f77bd1711c0394c6103950e973ea5939e577975263a8b368bd7877be"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.406001 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd0d1208f77bd1711c0394c6103950e973ea5939e577975263a8b368bd7877be" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.406062 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f34e-account-create-update-w5g78" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.412855 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qgs76" event={"ID":"9833ccd1-3ec0-487a-9ba3-b536a7f25829","Type":"ContainerStarted","Data":"b1d7416c65f885ca7a9f5fae131a4a5868e7716faec7f0727c62b2869e78ccee"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.414609 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-622c8" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.414615 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-622c8" event={"ID":"9c223cd8-b92c-470b-bf11-f0c41d352887","Type":"ContainerDied","Data":"42637e1c776ecdada5919c29af5fe759457dcd801110ea1b656304ee64cf9cae"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.415925 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42637e1c776ecdada5919c29af5fe759457dcd801110ea1b656304ee64cf9cae" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.416922 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-25q46" event={"ID":"590cb807-ad51-4176-a32c-0366c5d976c5","Type":"ContainerDied","Data":"3efe961e11b78b681de5df06f0131550b232523beda6ae7f9295a4ebf3bb441d"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.416961 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3efe961e11b78b681de5df06f0131550b232523beda6ae7f9295a4ebf3bb441d" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.417059 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-25q46" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.426583 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-23af-account-create-update-4vg5g" event={"ID":"fc65938f-ce7e-462d-bd24-b8265723c981","Type":"ContainerDied","Data":"a0f42e263521e642462b3cc44531a41d6c9d14905948bcfe626a339cc5bc63ab"} Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.426633 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0f42e263521e642462b3cc44531a41d6c9d14905948bcfe626a339cc5bc63ab" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.426751 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-23af-account-create-update-4vg5g" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.465829 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lsb5x"] Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.488518 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgfc7\" (UniqueName: \"kubernetes.io/projected/9c223cd8-b92c-470b-bf11-f0c41d352887-kube-api-access-vgfc7\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.488557 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c223cd8-b92c-470b-bf11-f0c41d352887-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.494953 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lsb5x"] Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.505437 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-qgs76" podStartSLOduration=2.900243524 podStartE2EDuration="8.505410033s" podCreationTimestamp="2025-11-27 11:56:07 +0000 UTC" firstStartedPulling="2025-11-27 11:56:09.303550647 +0000 UTC m=+1886.821869565" lastFinishedPulling="2025-11-27 11:56:14.908717156 +0000 UTC m=+1892.427036074" observedRunningTime="2025-11-27 11:56:15.458173023 +0000 UTC m=+1892.976491941" watchObservedRunningTime="2025-11-27 11:56:15.505410033 +0000 UTC m=+1893.023728961" Nov 27 11:56:15 crc kubenswrapper[4796]: I1127 11:56:15.579892 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" path="/var/lib/kubelet/pods/a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa/volumes" Nov 27 11:56:17 crc kubenswrapper[4796]: I1127 11:56:17.570054 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:56:17 crc kubenswrapper[4796]: E1127 11:56:17.570810 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:56:18 crc kubenswrapper[4796]: I1127 11:56:18.453059 4796 generic.go:334] "Generic (PLEG): container finished" podID="9833ccd1-3ec0-487a-9ba3-b536a7f25829" containerID="b1d7416c65f885ca7a9f5fae131a4a5868e7716faec7f0727c62b2869e78ccee" exitCode=0 Nov 27 11:56:18 crc kubenswrapper[4796]: I1127 11:56:18.453107 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qgs76" event={"ID":"9833ccd1-3ec0-487a-9ba3-b536a7f25829","Type":"ContainerDied","Data":"b1d7416c65f885ca7a9f5fae131a4a5868e7716faec7f0727c62b2869e78ccee"} Nov 27 11:56:19 crc kubenswrapper[4796]: I1127 11:56:19.784834 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:19 crc kubenswrapper[4796]: I1127 11:56:19.961608 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-config-data\") pod \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " Nov 27 11:56:19 crc kubenswrapper[4796]: I1127 11:56:19.961682 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z67bq\" (UniqueName: \"kubernetes.io/projected/9833ccd1-3ec0-487a-9ba3-b536a7f25829-kube-api-access-z67bq\") pod \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " Nov 27 11:56:19 crc kubenswrapper[4796]: I1127 11:56:19.961745 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-combined-ca-bundle\") pod \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\" (UID: \"9833ccd1-3ec0-487a-9ba3-b536a7f25829\") " Nov 27 11:56:19 crc kubenswrapper[4796]: I1127 11:56:19.966840 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9833ccd1-3ec0-487a-9ba3-b536a7f25829-kube-api-access-z67bq" (OuterVolumeSpecName: "kube-api-access-z67bq") pod "9833ccd1-3ec0-487a-9ba3-b536a7f25829" (UID: "9833ccd1-3ec0-487a-9ba3-b536a7f25829"). InnerVolumeSpecName "kube-api-access-z67bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:19 crc kubenswrapper[4796]: I1127 11:56:19.987798 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9833ccd1-3ec0-487a-9ba3-b536a7f25829" (UID: "9833ccd1-3ec0-487a-9ba3-b536a7f25829"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.010064 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-config-data" (OuterVolumeSpecName: "config-data") pod "9833ccd1-3ec0-487a-9ba3-b536a7f25829" (UID: "9833ccd1-3ec0-487a-9ba3-b536a7f25829"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.010425 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.074234 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.074284 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z67bq\" (UniqueName: \"kubernetes.io/projected/9833ccd1-3ec0-487a-9ba3-b536a7f25829-kube-api-access-z67bq\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.074294 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9833ccd1-3ec0-487a-9ba3-b536a7f25829-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.087965 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jpzwj"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.088280 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" podUID="d7630686-df2d-484c-81d3-769308420b18" containerName="dnsmasq-dns" containerID="cri-o://4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e" gracePeriod=10 Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.444236 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.469070 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qgs76" event={"ID":"9833ccd1-3ec0-487a-9ba3-b536a7f25829","Type":"ContainerDied","Data":"970c0af0a3dbd1ba42203571e0beb9e6fec7cec1c13ce7ff833f61fa7a0a302e"} Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.469136 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970c0af0a3dbd1ba42203571e0beb9e6fec7cec1c13ce7ff833f61fa7a0a302e" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.470636 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qgs76" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.473361 4796 generic.go:334] "Generic (PLEG): container finished" podID="d7630686-df2d-484c-81d3-769308420b18" containerID="4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e" exitCode=0 Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.473401 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" event={"ID":"d7630686-df2d-484c-81d3-769308420b18","Type":"ContainerDied","Data":"4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e"} Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.473427 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" event={"ID":"d7630686-df2d-484c-81d3-769308420b18","Type":"ContainerDied","Data":"1ba0998975266f85581e3b7f86c103289eae73b85f6b9d5f3dd53d1c92d8c970"} Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.473450 4796 scope.go:117] "RemoveContainer" containerID="4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.473588 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jpzwj" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.480554 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-dns-svc\") pod \"d7630686-df2d-484c-81d3-769308420b18\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.480679 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8cpp\" (UniqueName: \"kubernetes.io/projected/d7630686-df2d-484c-81d3-769308420b18-kube-api-access-c8cpp\") pod \"d7630686-df2d-484c-81d3-769308420b18\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.480758 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-nb\") pod \"d7630686-df2d-484c-81d3-769308420b18\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.480780 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-sb\") pod \"d7630686-df2d-484c-81d3-769308420b18\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.480833 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-config\") pod \"d7630686-df2d-484c-81d3-769308420b18\" (UID: \"d7630686-df2d-484c-81d3-769308420b18\") " Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.492866 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7630686-df2d-484c-81d3-769308420b18-kube-api-access-c8cpp" (OuterVolumeSpecName: "kube-api-access-c8cpp") pod "d7630686-df2d-484c-81d3-769308420b18" (UID: "d7630686-df2d-484c-81d3-769308420b18"). InnerVolumeSpecName "kube-api-access-c8cpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.507641 4796 scope.go:117] "RemoveContainer" containerID="d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.529554 4796 scope.go:117] "RemoveContainer" containerID="4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.530037 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e\": container with ID starting with 4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e not found: ID does not exist" containerID="4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.530072 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e"} err="failed to get container status \"4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e\": rpc error: code = NotFound desc = could not find container \"4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e\": container with ID starting with 4ba94ad19fa2bf4199570e24ad03e2b71b0be54f5fca8b782786d5bed752646e not found: ID does not exist" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.530092 4796 scope.go:117] "RemoveContainer" containerID="d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.530957 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e\": container with ID starting with d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e not found: ID does not exist" containerID="d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.530980 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e"} err="failed to get container status \"d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e\": rpc error: code = NotFound desc = could not find container \"d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e\": container with ID starting with d87ba319f28886504c95247194beaff41caa685ef4b87a406bee1df59d42446e not found: ID does not exist" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.537871 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-config" (OuterVolumeSpecName: "config") pod "d7630686-df2d-484c-81d3-769308420b18" (UID: "d7630686-df2d-484c-81d3-769308420b18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.541717 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7630686-df2d-484c-81d3-769308420b18" (UID: "d7630686-df2d-484c-81d3-769308420b18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.546033 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7630686-df2d-484c-81d3-769308420b18" (UID: "d7630686-df2d-484c-81d3-769308420b18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.548408 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7630686-df2d-484c-81d3-769308420b18" (UID: "d7630686-df2d-484c-81d3-769308420b18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.586798 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.587023 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8cpp\" (UniqueName: \"kubernetes.io/projected/d7630686-df2d-484c-81d3-769308420b18-kube-api-access-c8cpp\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.587090 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.587165 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.587216 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7630686-df2d-484c-81d3-769308420b18-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.639879 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-k5z54"] Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.643887 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e62fd9-c328-49cf-8f64-a61fb678a313" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.643973 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e62fd9-c328-49cf-8f64-a61fb678a313" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644079 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ac8842-ecad-4b5d-a7e9-6daa85540618" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.644144 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ac8842-ecad-4b5d-a7e9-6daa85540618" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644211 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" containerName="init" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.644340 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" containerName="init" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644425 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e84a3d5c-420f-437c-9ba9-901f550d54a7" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.644493 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e84a3d5c-420f-437c-9ba9-901f550d54a7" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644570 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="590cb807-ad51-4176-a32c-0366c5d976c5" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.644637 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="590cb807-ad51-4176-a32c-0366c5d976c5" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644707 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7630686-df2d-484c-81d3-769308420b18" containerName="dnsmasq-dns" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.644762 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7630686-df2d-484c-81d3-769308420b18" containerName="dnsmasq-dns" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644824 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc65938f-ce7e-462d-bd24-b8265723c981" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.644889 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc65938f-ce7e-462d-bd24-b8265723c981" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.644964 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c223cd8-b92c-470b-bf11-f0c41d352887" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645014 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c223cd8-b92c-470b-bf11-f0c41d352887" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.645065 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9938c442-847a-48f5-bcd4-7454e8a7c7e1" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645123 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9938c442-847a-48f5-bcd4-7454e8a7c7e1" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.645203 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7630686-df2d-484c-81d3-769308420b18" containerName="init" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645292 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7630686-df2d-484c-81d3-769308420b18" containerName="init" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.645356 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="430e7010-722c-4c22-9097-b65ed2975115" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645414 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="430e7010-722c-4c22-9097-b65ed2975115" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: E1127 11:56:20.645493 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9833ccd1-3ec0-487a-9ba3-b536a7f25829" containerName="keystone-db-sync" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645542 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9833ccd1-3ec0-487a-9ba3-b536a7f25829" containerName="keystone-db-sync" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645744 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ac8842-ecad-4b5d-a7e9-6daa85540618" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645810 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="590cb807-ad51-4176-a32c-0366c5d976c5" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645865 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e62fd9-c328-49cf-8f64-a61fb678a313" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645919 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7630686-df2d-484c-81d3-769308420b18" containerName="dnsmasq-dns" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.645983 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c1fc12-d2ae-467e-9be7-a6f16aa6e8fa" containerName="init" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.646053 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9833ccd1-3ec0-487a-9ba3-b536a7f25829" containerName="keystone-db-sync" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.646123 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c223cd8-b92c-470b-bf11-f0c41d352887" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.646174 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc65938f-ce7e-462d-bd24-b8265723c981" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.646225 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e84a3d5c-420f-437c-9ba9-901f550d54a7" containerName="mariadb-account-create-update" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.646295 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9938c442-847a-48f5-bcd4-7454e8a7c7e1" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.646351 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="430e7010-722c-4c22-9097-b65ed2975115" containerName="mariadb-database-create" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.647247 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.656074 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-k5z54"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.691076 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.691132 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.691174 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.691203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.691251 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27bph\" (UniqueName: \"kubernetes.io/projected/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-kube-api-access-27bph\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.691296 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-config\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.704278 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-lxj7c"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.705295 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.709698 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.709907 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.710096 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.710281 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.710490 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rc9hr" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.754061 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lxj7c"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.795910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-config\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.795983 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-fernet-keys\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796016 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-scripts\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796052 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-config-data\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796117 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jchp6\" (UniqueName: \"kubernetes.io/projected/e5200074-c987-416d-a17d-693d264d7ea8-kube-api-access-jchp6\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796148 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796191 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796249 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-combined-ca-bundle\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796296 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-credential-keys\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.796332 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27bph\" (UniqueName: \"kubernetes.io/projected/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-kube-api-access-27bph\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.801317 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.802800 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.802906 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-config\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.803006 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.803398 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.857188 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27bph\" (UniqueName: \"kubernetes.io/projected/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-kube-api-access-27bph\") pod \"dnsmasq-dns-bbf5cc879-k5z54\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.892406 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-km9nk"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.893656 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.897956 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-fernet-keys\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.897994 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-scripts\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.898024 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-config-data\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.898051 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jchp6\" (UniqueName: \"kubernetes.io/projected/e5200074-c987-416d-a17d-693d264d7ea8-kube-api-access-jchp6\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.898106 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-combined-ca-bundle\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.898132 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-credential-keys\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.899993 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.905237 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-fernet-keys\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.908870 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-credential-keys\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.909180 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-24wlq" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.909973 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-combined-ca-bundle\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.910296 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-config-data\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.911753 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-scripts\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.919253 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jpzwj"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.938007 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jpzwj"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.944669 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jchp6\" (UniqueName: \"kubernetes.io/projected/e5200074-c987-416d-a17d-693d264d7ea8-kube-api-access-jchp6\") pod \"keystone-bootstrap-lxj7c\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.952086 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-km9nk"] Nov 27 11:56:20 crc kubenswrapper[4796]: I1127 11:56:20.970699 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.000232 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-config-data\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.000326 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnbv4\" (UniqueName: \"kubernetes.io/projected/d920f619-249d-4266-ae4b-e13917ae78dc-kube-api-access-xnbv4\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.000399 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-combined-ca-bundle\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.027881 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-k5z54"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.068169 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.071303 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-jfd4m"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.072538 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.077837 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.078169 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-vmg9q" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.079033 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.083321 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-276z6"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.087094 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.090646 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-bbvsz"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.094179 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mj92m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.094409 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.094552 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.097960 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101419 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-config-data\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101454 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-combined-ca-bundle\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101483 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-combined-ca-bundle\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101511 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84tsj\" (UniqueName: \"kubernetes.io/projected/24d550fe-17da-4b68-8e62-b537d6c07314-kube-api-access-84tsj\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101534 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24d550fe-17da-4b68-8e62-b537d6c07314-etc-machine-id\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101565 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-config-data\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-combined-ca-bundle\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-db-sync-config-data\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101636 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-logs\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101655 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-scripts\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101680 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-config-data\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101701 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8sv9\" (UniqueName: \"kubernetes.io/projected/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-kube-api-access-f8sv9\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101717 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnbv4\" (UniqueName: \"kubernetes.io/projected/d920f619-249d-4266-ae4b-e13917ae78dc-kube-api-access-xnbv4\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.101734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-scripts\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.105354 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-combined-ca-bundle\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.109577 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-config-data\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.110017 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.110200 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.110653 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-v6cdm" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.121600 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dwbsz"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.122636 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.124679 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.124864 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9kkpt" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.144832 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnbv4\" (UniqueName: \"kubernetes.io/projected/d920f619-249d-4266-ae4b-e13917ae78dc-kube-api-access-xnbv4\") pod \"heat-db-sync-km9nk\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.144951 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jfd4m"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.159770 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bbvsz"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.186322 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.188144 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.189791 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.190507 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.203138 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-276z6"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208561 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-log-httpd\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208604 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkfq\" (UniqueName: \"kubernetes.io/projected/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-kube-api-access-sfkfq\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208636 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-combined-ca-bundle\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208656 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208677 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-db-sync-config-data\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-logs\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208720 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-scripts\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8sv9\" (UniqueName: \"kubernetes.io/projected/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-kube-api-access-f8sv9\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208764 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-config-data\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208788 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-scripts\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208807 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b5jt\" (UniqueName: \"kubernetes.io/projected/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-kube-api-access-6b5jt\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208838 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-config-data\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208858 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-combined-ca-bundle\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208884 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-combined-ca-bundle\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208909 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8whrf\" (UniqueName: \"kubernetes.io/projected/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-kube-api-access-8whrf\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208951 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208969 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-scripts\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.208990 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-config\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.209019 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-combined-ca-bundle\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.209035 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-run-httpd\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.209051 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-db-sync-config-data\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.209082 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84tsj\" (UniqueName: \"kubernetes.io/projected/24d550fe-17da-4b68-8e62-b537d6c07314-kube-api-access-84tsj\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.209118 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24d550fe-17da-4b68-8e62-b537d6c07314-etc-machine-id\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.209160 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-config-data\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.211084 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24d550fe-17da-4b68-8e62-b537d6c07314-etc-machine-id\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.211760 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-logs\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.215802 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-config-data\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.216212 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-combined-ca-bundle\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.219461 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-combined-ca-bundle\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.219659 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-config-data\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.223105 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-db-sync-config-data\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.225920 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-scripts\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.226924 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.235237 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-scripts\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.240372 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8sv9\" (UniqueName: \"kubernetes.io/projected/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-kube-api-access-f8sv9\") pod \"placement-db-sync-276z6\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.240447 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-mkwn4"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.248837 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.249809 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dwbsz"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.251921 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84tsj\" (UniqueName: \"kubernetes.io/projected/24d550fe-17da-4b68-8e62-b537d6c07314-kube-api-access-84tsj\") pod \"cinder-db-sync-jfd4m\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.269379 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-mkwn4"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311307 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-config-data\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311618 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b5jt\" (UniqueName: \"kubernetes.io/projected/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-kube-api-access-6b5jt\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311641 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nl5q\" (UniqueName: \"kubernetes.io/projected/680cc550-1bfc-4a40-97e5-72067cb9efb7-kube-api-access-7nl5q\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311670 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-combined-ca-bundle\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311690 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311704 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311721 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8whrf\" (UniqueName: \"kubernetes.io/projected/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-kube-api-access-8whrf\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311752 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-scripts\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-config\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-combined-ca-bundle\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311802 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-run-httpd\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311817 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-db-sync-config-data\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311865 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311885 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-log-httpd\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311901 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkfq\" (UniqueName: \"kubernetes.io/projected/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-kube-api-access-sfkfq\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311922 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311943 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.311962 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-config\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.313277 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-log-httpd\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.313496 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-run-httpd\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.318923 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-scripts\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.319392 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-config-data\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.319760 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.322403 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-db-sync-config-data\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.322525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-combined-ca-bundle\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.323427 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.323567 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-combined-ca-bundle\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.323841 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-config\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.330552 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b5jt\" (UniqueName: \"kubernetes.io/projected/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-kube-api-access-6b5jt\") pod \"ceilometer-0\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.331748 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8whrf\" (UniqueName: \"kubernetes.io/projected/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-kube-api-access-8whrf\") pod \"neutron-db-sync-bbvsz\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.334640 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkfq\" (UniqueName: \"kubernetes.io/projected/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-kube-api-access-sfkfq\") pod \"barbican-db-sync-dwbsz\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.350207 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-km9nk" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.414563 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.414635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-config\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.414694 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nl5q\" (UniqueName: \"kubernetes.io/projected/680cc550-1bfc-4a40-97e5-72067cb9efb7-kube-api-access-7nl5q\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.414728 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.414742 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.414806 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.415572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.416151 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.416575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.416590 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-config\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.417151 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.438849 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nl5q\" (UniqueName: \"kubernetes.io/projected/680cc550-1bfc-4a40-97e5-72067cb9efb7-kube-api-access-7nl5q\") pod \"dnsmasq-dns-56df8fb6b7-mkwn4\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.443795 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.476135 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-276z6" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.506353 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.529178 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.548656 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.553608 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lxj7c"] Nov 27 11:56:21 crc kubenswrapper[4796]: W1127 11:56:21.589439 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5200074_c987_416d_a17d_693d264d7ea8.slice/crio-da627b54435d3d56780e2833e771d033e57cdd03b4ba824539c731de3e6bdd9b WatchSource:0}: Error finding container da627b54435d3d56780e2833e771d033e57cdd03b4ba824539c731de3e6bdd9b: Status 404 returned error can't find the container with id da627b54435d3d56780e2833e771d033e57cdd03b4ba824539c731de3e6bdd9b Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.589842 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.595722 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7630686-df2d-484c-81d3-769308420b18" path="/var/lib/kubelet/pods/d7630686-df2d-484c-81d3-769308420b18/volumes" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.636597 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-k5z54"] Nov 27 11:56:21 crc kubenswrapper[4796]: W1127 11:56:21.672027 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6779f8f8_0367_431a_8e8c_a5e70ba0b93c.slice/crio-0f7bcd878fb36e7394493b14ba205c2fd9ef21f0b47a7ea5e4b68618f3601f0d WatchSource:0}: Error finding container 0f7bcd878fb36e7394493b14ba205c2fd9ef21f0b47a7ea5e4b68618f3601f0d: Status 404 returned error can't find the container with id 0f7bcd878fb36e7394493b14ba205c2fd9ef21f0b47a7ea5e4b68618f3601f0d Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.897256 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.905762 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.912355 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.912681 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.913010 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.925487 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gbfzq" Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.955137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.985823 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:21 crc kubenswrapper[4796]: I1127 11:56:21.987243 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.011937 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.013401 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.022813 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.032863 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-km9nk"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051427 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051469 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpk2h\" (UniqueName: \"kubernetes.io/projected/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-kube-api-access-xpk2h\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051493 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051530 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051554 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051595 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051628 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-logs\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.051647 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153021 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153064 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153110 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153138 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153167 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153186 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-logs\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153210 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153232 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153257 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153294 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153310 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpk2h\" (UniqueName: \"kubernetes.io/projected/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-kube-api-access-xpk2h\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx8f8\" (UniqueName: \"kubernetes.io/projected/f44490bf-8c3c-4252-996f-a350bf84280e-kube-api-access-vx8f8\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153345 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153365 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153387 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.153413 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.155539 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.155758 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-logs\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.158647 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.159145 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.162668 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.166087 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.173317 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.183116 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpk2h\" (UniqueName: \"kubernetes.io/projected/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-kube-api-access-xpk2h\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.219238 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259371 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259417 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259478 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259529 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259571 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259618 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.259682 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx8f8\" (UniqueName: \"kubernetes.io/projected/f44490bf-8c3c-4252-996f-a350bf84280e-kube-api-access-vx8f8\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.260354 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-logs\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.261235 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.261441 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.265699 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.265817 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.265956 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-276z6"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.266962 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.280244 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jfd4m"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.282000 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx8f8\" (UniqueName: \"kubernetes.io/projected/f44490bf-8c3c-4252-996f-a350bf84280e-kube-api-access-vx8f8\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.285478 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.288501 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: W1127 11:56:22.403339 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4924cd5e_791b_4312_b9bc_ad1167ec5fe4.slice/crio-22655bb8f0ef7bdd64bd0bff96072898d987a46bb6bc15b89094cf1e687fea42 WatchSource:0}: Error finding container 22655bb8f0ef7bdd64bd0bff96072898d987a46bb6bc15b89094cf1e687fea42: Status 404 returned error can't find the container with id 22655bb8f0ef7bdd64bd0bff96072898d987a46bb6bc15b89094cf1e687fea42 Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.403552 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bbvsz"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.410285 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.437959 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.502153 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dwbsz"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.512734 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfd4m" event={"ID":"24d550fe-17da-4b68-8e62-b537d6c07314","Type":"ContainerStarted","Data":"f99347f3a40580cf78a3f7bc7d85f436fe32bf185d01c56760bc7e5111a264a2"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.518045 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lxj7c" event={"ID":"e5200074-c987-416d-a17d-693d264d7ea8","Type":"ContainerStarted","Data":"7f6e3ea3dafac0198575955c4a149c4463bc8493e9f3101066b2fe281241530e"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.518114 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lxj7c" event={"ID":"e5200074-c987-416d-a17d-693d264d7ea8","Type":"ContainerStarted","Data":"da627b54435d3d56780e2833e771d033e57cdd03b4ba824539c731de3e6bdd9b"} Nov 27 11:56:22 crc kubenswrapper[4796]: W1127 11:56:22.518322 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8d6ccfb_6b49_4e63_b4fe_2e061f4fc6e3.slice/crio-6f9fcfb451eac04360b1763d55bf189516b735156163ab9293162721cded688a WatchSource:0}: Error finding container 6f9fcfb451eac04360b1763d55bf189516b735156163ab9293162721cded688a: Status 404 returned error can't find the container with id 6f9fcfb451eac04360b1763d55bf189516b735156163ab9293162721cded688a Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.527261 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.549453 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-276z6" event={"ID":"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8","Type":"ContainerStarted","Data":"b049de2f72cdebded9e4c5f460de98593c5ca9e5dc60f8abc8c18aebf8dd583d"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.556480 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-lxj7c" podStartSLOduration=2.556461664 podStartE2EDuration="2.556461664s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:22.537704143 +0000 UTC m=+1900.056023071" watchObservedRunningTime="2025-11-27 11:56:22.556461664 +0000 UTC m=+1900.074780582" Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.567176 4796 generic.go:334] "Generic (PLEG): container finished" podID="6779f8f8-0367-431a-8e8c-a5e70ba0b93c" containerID="a89d1a09a0f4cfd0ca8b115c61308a0db2e9ea6f08806a15fd8bba76aa08d807" exitCode=0 Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.567557 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" event={"ID":"6779f8f8-0367-431a-8e8c-a5e70ba0b93c","Type":"ContainerDied","Data":"a89d1a09a0f4cfd0ca8b115c61308a0db2e9ea6f08806a15fd8bba76aa08d807"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.567655 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" event={"ID":"6779f8f8-0367-431a-8e8c-a5e70ba0b93c","Type":"ContainerStarted","Data":"0f7bcd878fb36e7394493b14ba205c2fd9ef21f0b47a7ea5e4b68618f3601f0d"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.571281 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-km9nk" event={"ID":"d920f619-249d-4266-ae4b-e13917ae78dc","Type":"ContainerStarted","Data":"bbadb8bb7ea5a4e48d9e5ac79e58f16d5c4da96e50285d010ceca03e437d4382"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.572348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bbvsz" event={"ID":"4924cd5e-791b-4312-b9bc-ad1167ec5fe4","Type":"ContainerStarted","Data":"22655bb8f0ef7bdd64bd0bff96072898d987a46bb6bc15b89094cf1e687fea42"} Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.635676 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.713562 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.751426 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-mkwn4"] Nov 27 11:56:22 crc kubenswrapper[4796]: I1127 11:56:22.799180 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.078567 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.190013 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-sb\") pod \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.190106 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-swift-storage-0\") pod \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.190151 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-svc\") pod \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.190214 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-config\") pod \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.190312 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-nb\") pod \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.190341 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27bph\" (UniqueName: \"kubernetes.io/projected/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-kube-api-access-27bph\") pod \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\" (UID: \"6779f8f8-0367-431a-8e8c-a5e70ba0b93c\") " Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.200724 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.200998 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-kube-api-access-27bph" (OuterVolumeSpecName: "kube-api-access-27bph") pod "6779f8f8-0367-431a-8e8c-a5e70ba0b93c" (UID: "6779f8f8-0367-431a-8e8c-a5e70ba0b93c"). InnerVolumeSpecName "kube-api-access-27bph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.218460 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6779f8f8-0367-431a-8e8c-a5e70ba0b93c" (UID: "6779f8f8-0367-431a-8e8c-a5e70ba0b93c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.220029 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6779f8f8-0367-431a-8e8c-a5e70ba0b93c" (UID: "6779f8f8-0367-431a-8e8c-a5e70ba0b93c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.220378 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6779f8f8-0367-431a-8e8c-a5e70ba0b93c" (UID: "6779f8f8-0367-431a-8e8c-a5e70ba0b93c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.221360 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6779f8f8-0367-431a-8e8c-a5e70ba0b93c" (UID: "6779f8f8-0367-431a-8e8c-a5e70ba0b93c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.232414 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-config" (OuterVolumeSpecName: "config") pod "6779f8f8-0367-431a-8e8c-a5e70ba0b93c" (UID: "6779f8f8-0367-431a-8e8c-a5e70ba0b93c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.297201 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.297235 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27bph\" (UniqueName: \"kubernetes.io/projected/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-kube-api-access-27bph\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.297247 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.297255 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.297277 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.297286 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6779f8f8-0367-431a-8e8c-a5e70ba0b93c-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.301125 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.587193 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" event={"ID":"6779f8f8-0367-431a-8e8c-a5e70ba0b93c","Type":"ContainerDied","Data":"0f7bcd878fb36e7394493b14ba205c2fd9ef21f0b47a7ea5e4b68618f3601f0d"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.587245 4796 scope.go:117] "RemoveContainer" containerID="a89d1a09a0f4cfd0ca8b115c61308a0db2e9ea6f08806a15fd8bba76aa08d807" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.587376 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-k5z54" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.593570 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bbvsz" event={"ID":"4924cd5e-791b-4312-b9bc-ad1167ec5fe4","Type":"ContainerStarted","Data":"3817b175de387b9b45e2967aac4ebb34c0f8e5a6a6a82648b0bacf995423d79a"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.595768 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dwbsz" event={"ID":"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3","Type":"ContainerStarted","Data":"6f9fcfb451eac04360b1763d55bf189516b735156163ab9293162721cded688a"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.602794 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerStarted","Data":"85614f297a6402221e4cd7ce5427483143e7b5667372fc4ddf917e3035313915"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.605282 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa","Type":"ContainerStarted","Data":"29d566cbd4f1da2035bd7a15c1435ed221b83852abe13f88c180d242c1b313f9"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.613903 4796 generic.go:334] "Generic (PLEG): container finished" podID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerID="2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054" exitCode=0 Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.613971 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" event={"ID":"680cc550-1bfc-4a40-97e5-72067cb9efb7","Type":"ContainerDied","Data":"2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.613996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" event={"ID":"680cc550-1bfc-4a40-97e5-72067cb9efb7","Type":"ContainerStarted","Data":"697e421b63756cae1b525b76c4b1f66075f24b7d15a817f550581400aff96eec"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.623155 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f44490bf-8c3c-4252-996f-a350bf84280e","Type":"ContainerStarted","Data":"1f337f136bf0b648bcb41d8ad662daf1c0d2014d19f90a91a7244050c6e8608a"} Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.801832 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-bbvsz" podStartSLOduration=3.801811115 podStartE2EDuration="3.801811115s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:23.786282832 +0000 UTC m=+1901.304601750" watchObservedRunningTime="2025-11-27 11:56:23.801811115 +0000 UTC m=+1901.320130033" Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.843934 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-k5z54"] Nov 27 11:56:23 crc kubenswrapper[4796]: I1127 11:56:23.874929 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-k5z54"] Nov 27 11:56:24 crc kubenswrapper[4796]: I1127 11:56:24.664622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f44490bf-8c3c-4252-996f-a350bf84280e","Type":"ContainerStarted","Data":"4d78e6316eb3cb08efb46fa4783110ed5194c072a8b25766da32bea2abe11914"} Nov 27 11:56:24 crc kubenswrapper[4796]: I1127 11:56:24.678388 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa","Type":"ContainerStarted","Data":"5a0fe5cecb60e802ab70dae2077c32f1566c0384e34c7b153e499573d7a2a9e4"} Nov 27 11:56:24 crc kubenswrapper[4796]: I1127 11:56:24.684934 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" event={"ID":"680cc550-1bfc-4a40-97e5-72067cb9efb7","Type":"ContainerStarted","Data":"eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b"} Nov 27 11:56:24 crc kubenswrapper[4796]: I1127 11:56:24.684973 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:24 crc kubenswrapper[4796]: I1127 11:56:24.711210 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" podStartSLOduration=3.711193155 podStartE2EDuration="3.711193155s" podCreationTimestamp="2025-11-27 11:56:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:24.706599172 +0000 UTC m=+1902.224918090" watchObservedRunningTime="2025-11-27 11:56:24.711193155 +0000 UTC m=+1902.229512073" Nov 27 11:56:25 crc kubenswrapper[4796]: I1127 11:56:25.586369 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6779f8f8-0367-431a-8e8c-a5e70ba0b93c" path="/var/lib/kubelet/pods/6779f8f8-0367-431a-8e8c-a5e70ba0b93c/volumes" Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.702565 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa","Type":"ContainerStarted","Data":"85de6210484f7d1c2d353c58fd370c7919b81ed0d3f5813709e7cd1fa1dc3b27"} Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.702652 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-log" containerID="cri-o://5a0fe5cecb60e802ab70dae2077c32f1566c0384e34c7b153e499573d7a2a9e4" gracePeriod=30 Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.702879 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-httpd" containerID="cri-o://85de6210484f7d1c2d353c58fd370c7919b81ed0d3f5813709e7cd1fa1dc3b27" gracePeriod=30 Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.706347 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f44490bf-8c3c-4252-996f-a350bf84280e","Type":"ContainerStarted","Data":"fce65c12961dc5cfed4b9e193ed9605083c5d9715b1431091bf835a0ea626314"} Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.706458 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-log" containerID="cri-o://4d78e6316eb3cb08efb46fa4783110ed5194c072a8b25766da32bea2abe11914" gracePeriod=30 Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.706531 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-httpd" containerID="cri-o://fce65c12961dc5cfed4b9e193ed9605083c5d9715b1431091bf835a0ea626314" gracePeriod=30 Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.727834 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.727820503 podStartE2EDuration="6.727820503s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:26.722881331 +0000 UTC m=+1904.241200249" watchObservedRunningTime="2025-11-27 11:56:26.727820503 +0000 UTC m=+1904.246139421" Nov 27 11:56:26 crc kubenswrapper[4796]: I1127 11:56:26.740848 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.740829939 podStartE2EDuration="6.740829939s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:26.739232847 +0000 UTC m=+1904.257551765" watchObservedRunningTime="2025-11-27 11:56:26.740829939 +0000 UTC m=+1904.259148857" Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.718586 4796 generic.go:334] "Generic (PLEG): container finished" podID="e5200074-c987-416d-a17d-693d264d7ea8" containerID="7f6e3ea3dafac0198575955c4a149c4463bc8493e9f3101066b2fe281241530e" exitCode=0 Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.718667 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lxj7c" event={"ID":"e5200074-c987-416d-a17d-693d264d7ea8","Type":"ContainerDied","Data":"7f6e3ea3dafac0198575955c4a149c4463bc8493e9f3101066b2fe281241530e"} Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.722543 4796 generic.go:334] "Generic (PLEG): container finished" podID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerID="85de6210484f7d1c2d353c58fd370c7919b81ed0d3f5813709e7cd1fa1dc3b27" exitCode=0 Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.722907 4796 generic.go:334] "Generic (PLEG): container finished" podID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerID="5a0fe5cecb60e802ab70dae2077c32f1566c0384e34c7b153e499573d7a2a9e4" exitCode=143 Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.722613 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa","Type":"ContainerDied","Data":"85de6210484f7d1c2d353c58fd370c7919b81ed0d3f5813709e7cd1fa1dc3b27"} Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.723052 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa","Type":"ContainerDied","Data":"5a0fe5cecb60e802ab70dae2077c32f1566c0384e34c7b153e499573d7a2a9e4"} Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.726318 4796 generic.go:334] "Generic (PLEG): container finished" podID="f44490bf-8c3c-4252-996f-a350bf84280e" containerID="fce65c12961dc5cfed4b9e193ed9605083c5d9715b1431091bf835a0ea626314" exitCode=0 Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.726344 4796 generic.go:334] "Generic (PLEG): container finished" podID="f44490bf-8c3c-4252-996f-a350bf84280e" containerID="4d78e6316eb3cb08efb46fa4783110ed5194c072a8b25766da32bea2abe11914" exitCode=143 Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.726360 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f44490bf-8c3c-4252-996f-a350bf84280e","Type":"ContainerDied","Data":"fce65c12961dc5cfed4b9e193ed9605083c5d9715b1431091bf835a0ea626314"} Nov 27 11:56:27 crc kubenswrapper[4796]: I1127 11:56:27.726381 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f44490bf-8c3c-4252-996f-a350bf84280e","Type":"ContainerDied","Data":"4d78e6316eb3cb08efb46fa4783110ed5194c072a8b25766da32bea2abe11914"} Nov 27 11:56:31 crc kubenswrapper[4796]: I1127 11:56:31.592094 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:56:31 crc kubenswrapper[4796]: I1127 11:56:31.659585 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ngtn7"] Nov 27 11:56:31 crc kubenswrapper[4796]: I1127 11:56:31.659855 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" containerID="cri-o://34355bf4aac94f65b33f4030a233e523b4d7e7a9d1bb86f41f18d35a702e77bf" gracePeriod=10 Nov 27 11:56:32 crc kubenswrapper[4796]: I1127 11:56:32.569631 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:56:32 crc kubenswrapper[4796]: E1127 11:56:32.570228 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:56:32 crc kubenswrapper[4796]: I1127 11:56:32.793756 4796 generic.go:334] "Generic (PLEG): container finished" podID="b5504fa9-43f7-442d-a025-245928896378" containerID="34355bf4aac94f65b33f4030a233e523b4d7e7a9d1bb86f41f18d35a702e77bf" exitCode=0 Nov 27 11:56:32 crc kubenswrapper[4796]: I1127 11:56:32.793808 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" event={"ID":"b5504fa9-43f7-442d-a025-245928896378","Type":"ContainerDied","Data":"34355bf4aac94f65b33f4030a233e523b4d7e7a9d1bb86f41f18d35a702e77bf"} Nov 27 11:56:35 crc kubenswrapper[4796]: I1127 11:56:35.009753 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: connect: connection refused" Nov 27 11:56:37 crc kubenswrapper[4796]: E1127 11:56:37.015635 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 27 11:56:37 crc kubenswrapper[4796]: E1127 11:56:37.015831 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f8sv9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-276z6_openstack(23ffa4b0-e22e-4714-9b05-96bd3fffbfa8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:56:37 crc kubenswrapper[4796]: E1127 11:56:37.017086 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-276z6" podUID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" Nov 27 11:56:37 crc kubenswrapper[4796]: E1127 11:56:37.846973 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-276z6" podUID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.535515 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.647704 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.647869 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-httpd-run\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.647948 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-scripts\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.647994 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-config-data\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.648105 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx8f8\" (UniqueName: \"kubernetes.io/projected/f44490bf-8c3c-4252-996f-a350bf84280e-kube-api-access-vx8f8\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.648228 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-combined-ca-bundle\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.648291 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-internal-tls-certs\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.648342 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-logs\") pod \"f44490bf-8c3c-4252-996f-a350bf84280e\" (UID: \"f44490bf-8c3c-4252-996f-a350bf84280e\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.648427 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.648916 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.650033 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-logs" (OuterVolumeSpecName: "logs") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.653905 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f44490bf-8c3c-4252-996f-a350bf84280e-kube-api-access-vx8f8" (OuterVolumeSpecName: "kube-api-access-vx8f8") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "kube-api-access-vx8f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.655172 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.655615 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-scripts" (OuterVolumeSpecName: "scripts") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.688146 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.693500 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-config-data" (OuterVolumeSpecName: "config-data") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.703680 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f44490bf-8c3c-4252-996f-a350bf84280e" (UID: "f44490bf-8c3c-4252-996f-a350bf84280e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750371 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750402 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750416 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx8f8\" (UniqueName: \"kubernetes.io/projected/f44490bf-8c3c-4252-996f-a350bf84280e-kube-api-access-vx8f8\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750427 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750435 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44490bf-8c3c-4252-996f-a350bf84280e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750491 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44490bf-8c3c-4252-996f-a350bf84280e-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.750521 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.765798 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.809170 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.818760 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.818951 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dfh698hb7h645h5b6h5ddh54dh557h597h566h6chf6h56hf8hcfhd5h565h7bh647h5b5h694h5f7h57bh5c7h66fh5b8h558h548hb5h5cbh7dh9cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6b5jt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(13ee34c5-1762-4002-ba1e-63f3e1c90e6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.852109 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.899826 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa","Type":"ContainerDied","Data":"29d566cbd4f1da2035bd7a15c1435ed221b83852abe13f88c180d242c1b313f9"} Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.899885 4796 scope.go:117] "RemoveContainer" containerID="85de6210484f7d1c2d353c58fd370c7919b81ed0d3f5813709e7cd1fa1dc3b27" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.899845 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.903962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f44490bf-8c3c-4252-996f-a350bf84280e","Type":"ContainerDied","Data":"1f337f136bf0b648bcb41d8ad662daf1c0d2014d19f90a91a7244050c6e8608a"} Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.904015 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.939040 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953298 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-logs\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953433 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-combined-ca-bundle\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953507 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-config-data\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953539 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-httpd-run\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953594 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-scripts\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953627 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-public-tls-certs\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953701 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpk2h\" (UniqueName: \"kubernetes.io/projected/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-kube-api-access-xpk2h\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.953742 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\" (UID: \"ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa\") " Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.959948 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.960044 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-logs" (OuterVolumeSpecName: "logs") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.961234 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.963471 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.963492 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-kube-api-access-xpk2h" (OuterVolumeSpecName: "kube-api-access-xpk2h") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "kube-api-access-xpk2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.964171 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-scripts" (OuterVolumeSpecName: "scripts") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.976783 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.977176 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-log" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977196 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-log" Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.977204 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6779f8f8-0367-431a-8e8c-a5e70ba0b93c" containerName="init" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977211 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6779f8f8-0367-431a-8e8c-a5e70ba0b93c" containerName="init" Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.977226 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-httpd" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977234 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-httpd" Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.977244 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-log" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977250 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-log" Nov 27 11:56:39 crc kubenswrapper[4796]: E1127 11:56:39.977283 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-httpd" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977289 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-httpd" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977450 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-log" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977464 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-log" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977475 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" containerName="glance-httpd" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977488 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6779f8f8-0367-431a-8e8c-a5e70ba0b93c" containerName="init" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.977500 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" containerName="glance-httpd" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.978387 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.981200 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.981338 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 11:56:39 crc kubenswrapper[4796]: I1127 11:56:39.996778 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.046762 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.051608 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-config-data" (OuterVolumeSpecName: "config-data") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055592 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055624 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055634 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055644 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055652 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055661 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpk2h\" (UniqueName: \"kubernetes.io/projected/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-kube-api-access-xpk2h\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.055685 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.072339 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" (UID: "ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.078393 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.158258 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.158326 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.158571 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-logs\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.158977 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.159171 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.159239 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fshkp\" (UniqueName: \"kubernetes.io/projected/453af780-fb5a-4048-b9fd-2ce7a497916c-kube-api-access-fshkp\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.159441 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.159480 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.159585 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.159608 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.242029 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.253684 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261711 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261774 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261846 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261884 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-logs\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261956 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.261995 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.262022 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fshkp\" (UniqueName: \"kubernetes.io/projected/453af780-fb5a-4048-b9fd-2ce7a497916c-kube-api-access-fshkp\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.262511 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.263402 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-logs\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.263894 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.265617 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.267107 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.268171 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.270764 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.269832 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.274430 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.284659 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.284849 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.289214 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fshkp\" (UniqueName: \"kubernetes.io/projected/453af780-fb5a-4048-b9fd-2ce7a497916c-kube-api-access-fshkp\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.320940 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.343958 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.363519 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.363660 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.363718 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-config-data\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.363784 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr9xs\" (UniqueName: \"kubernetes.io/projected/2960ce96-6701-446e-a800-6f5f9333a322-kube-api-access-pr9xs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.364086 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.364205 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-scripts\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.364390 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-logs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.364452 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467013 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr9xs\" (UniqueName: \"kubernetes.io/projected/2960ce96-6701-446e-a800-6f5f9333a322-kube-api-access-pr9xs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467092 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467127 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-scripts\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467200 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-logs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467232 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467280 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467338 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467388 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-config-data\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.467937 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-logs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.469179 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.469732 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.475022 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.476480 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-config-data\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.477548 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.477715 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-scripts\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.490365 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr9xs\" (UniqueName: \"kubernetes.io/projected/2960ce96-6701-446e-a800-6f5f9333a322-kube-api-access-pr9xs\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.501391 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " pod="openstack/glance-default-external-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.610857 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:56:40 crc kubenswrapper[4796]: I1127 11:56:40.620539 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:56:41 crc kubenswrapper[4796]: I1127 11:56:41.579164 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa" path="/var/lib/kubelet/pods/ef1e7d0d-b56a-4d8f-abd6-06e5b52aa6fa/volumes" Nov 27 11:56:41 crc kubenswrapper[4796]: I1127 11:56:41.581044 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f44490bf-8c3c-4252-996f-a350bf84280e" path="/var/lib/kubelet/pods/f44490bf-8c3c-4252-996f-a350bf84280e/volumes" Nov 27 11:56:45 crc kubenswrapper[4796]: I1127 11:56:45.010740 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Nov 27 11:56:46 crc kubenswrapper[4796]: I1127 11:56:46.569000 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:56:46 crc kubenswrapper[4796]: E1127 11:56:46.569382 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:56:46 crc kubenswrapper[4796]: I1127 11:56:46.987857 4796 generic.go:334] "Generic (PLEG): container finished" podID="4924cd5e-791b-4312-b9bc-ad1167ec5fe4" containerID="3817b175de387b9b45e2967aac4ebb34c0f8e5a6a6a82648b0bacf995423d79a" exitCode=0 Nov 27 11:56:46 crc kubenswrapper[4796]: I1127 11:56:46.987921 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bbvsz" event={"ID":"4924cd5e-791b-4312-b9bc-ad1167ec5fe4","Type":"ContainerDied","Data":"3817b175de387b9b45e2967aac4ebb34c0f8e5a6a6a82648b0bacf995423d79a"} Nov 27 11:56:47 crc kubenswrapper[4796]: E1127 11:56:47.212821 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Nov 27 11:56:47 crc kubenswrapper[4796]: E1127 11:56:47.213006 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xnbv4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-km9nk_openstack(d920f619-249d-4266-ae4b-e13917ae78dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:56:47 crc kubenswrapper[4796]: E1127 11:56:47.214346 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-km9nk" podUID="d920f619-249d-4266-ae4b-e13917ae78dc" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.301317 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.321942 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.407910 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-credential-keys\") pod \"e5200074-c987-416d-a17d-693d264d7ea8\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.407996 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-scripts\") pod \"e5200074-c987-416d-a17d-693d264d7ea8\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408030 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-config-data\") pod \"e5200074-c987-416d-a17d-693d264d7ea8\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408121 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-fernet-keys\") pod \"e5200074-c987-416d-a17d-693d264d7ea8\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408157 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-config\") pod \"b5504fa9-43f7-442d-a025-245928896378\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408195 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-combined-ca-bundle\") pod \"e5200074-c987-416d-a17d-693d264d7ea8\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408293 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-swift-storage-0\") pod \"b5504fa9-43f7-442d-a025-245928896378\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408315 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jchp6\" (UniqueName: \"kubernetes.io/projected/e5200074-c987-416d-a17d-693d264d7ea8-kube-api-access-jchp6\") pod \"e5200074-c987-416d-a17d-693d264d7ea8\" (UID: \"e5200074-c987-416d-a17d-693d264d7ea8\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408360 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqjrk\" (UniqueName: \"kubernetes.io/projected/b5504fa9-43f7-442d-a025-245928896378-kube-api-access-dqjrk\") pod \"b5504fa9-43f7-442d-a025-245928896378\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408394 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-nb\") pod \"b5504fa9-43f7-442d-a025-245928896378\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.408416 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-sb\") pod \"b5504fa9-43f7-442d-a025-245928896378\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.409104 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-svc\") pod \"b5504fa9-43f7-442d-a025-245928896378\" (UID: \"b5504fa9-43f7-442d-a025-245928896378\") " Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.417693 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5504fa9-43f7-442d-a025-245928896378-kube-api-access-dqjrk" (OuterVolumeSpecName: "kube-api-access-dqjrk") pod "b5504fa9-43f7-442d-a025-245928896378" (UID: "b5504fa9-43f7-442d-a025-245928896378"). InnerVolumeSpecName "kube-api-access-dqjrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.418236 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e5200074-c987-416d-a17d-693d264d7ea8" (UID: "e5200074-c987-416d-a17d-693d264d7ea8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.423979 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5200074-c987-416d-a17d-693d264d7ea8-kube-api-access-jchp6" (OuterVolumeSpecName: "kube-api-access-jchp6") pod "e5200074-c987-416d-a17d-693d264d7ea8" (UID: "e5200074-c987-416d-a17d-693d264d7ea8"). InnerVolumeSpecName "kube-api-access-jchp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.427544 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-scripts" (OuterVolumeSpecName: "scripts") pod "e5200074-c987-416d-a17d-693d264d7ea8" (UID: "e5200074-c987-416d-a17d-693d264d7ea8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.433496 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e5200074-c987-416d-a17d-693d264d7ea8" (UID: "e5200074-c987-416d-a17d-693d264d7ea8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.442543 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-config-data" (OuterVolumeSpecName: "config-data") pod "e5200074-c987-416d-a17d-693d264d7ea8" (UID: "e5200074-c987-416d-a17d-693d264d7ea8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.460663 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5504fa9-43f7-442d-a025-245928896378" (UID: "b5504fa9-43f7-442d-a025-245928896378"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.461978 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5200074-c987-416d-a17d-693d264d7ea8" (UID: "e5200074-c987-416d-a17d-693d264d7ea8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.463371 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-config" (OuterVolumeSpecName: "config") pod "b5504fa9-43f7-442d-a025-245928896378" (UID: "b5504fa9-43f7-442d-a025-245928896378"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.466256 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5504fa9-43f7-442d-a025-245928896378" (UID: "b5504fa9-43f7-442d-a025-245928896378"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.478231 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5504fa9-43f7-442d-a025-245928896378" (UID: "b5504fa9-43f7-442d-a025-245928896378"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.479774 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5504fa9-43f7-442d-a025-245928896378" (UID: "b5504fa9-43f7-442d-a025-245928896378"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512355 4796 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512409 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512430 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512448 4796 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512474 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512490 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5200074-c987-416d-a17d-693d264d7ea8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512507 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512525 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jchp6\" (UniqueName: \"kubernetes.io/projected/e5200074-c987-416d-a17d-693d264d7ea8-kube-api-access-jchp6\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512549 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqjrk\" (UniqueName: \"kubernetes.io/projected/b5504fa9-43f7-442d-a025-245928896378-kube-api-access-dqjrk\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512566 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512583 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:47 crc kubenswrapper[4796]: I1127 11:56:47.512598 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5504fa9-43f7-442d-a025-245928896378-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.001174 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lxj7c" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.001538 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lxj7c" event={"ID":"e5200074-c987-416d-a17d-693d264d7ea8","Type":"ContainerDied","Data":"da627b54435d3d56780e2833e771d033e57cdd03b4ba824539c731de3e6bdd9b"} Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.001607 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da627b54435d3d56780e2833e771d033e57cdd03b4ba824539c731de3e6bdd9b" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.004512 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" event={"ID":"b5504fa9-43f7-442d-a025-245928896378","Type":"ContainerDied","Data":"b43aa1c084be18beb0c7d8c6ab9e127070bc6157d32a78cce808f9aa3a62392d"} Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.004589 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.005984 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-km9nk" podUID="d920f619-249d-4266-ae4b-e13917ae78dc" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.041963 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ngtn7"] Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.049182 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-ngtn7"] Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.385457 4796 scope.go:117] "RemoveContainer" containerID="5a0fe5cecb60e802ab70dae2077c32f1566c0384e34c7b153e499573d7a2a9e4" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.438376 4796 scope.go:117] "RemoveContainer" containerID="fce65c12961dc5cfed4b9e193ed9605083c5d9715b1431091bf835a0ea626314" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.455007 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.455178 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-84tsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-jfd4m_openstack(24d550fe-17da-4b68-8e62-b537d6c07314): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.456507 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-jfd4m" podUID="24d550fe-17da-4b68-8e62-b537d6c07314" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.476929 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-lxj7c"] Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.483924 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-lxj7c"] Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.527098 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583073 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-svvgz"] Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.583498 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4924cd5e-791b-4312-b9bc-ad1167ec5fe4" containerName="neutron-db-sync" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583510 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4924cd5e-791b-4312-b9bc-ad1167ec5fe4" containerName="neutron-db-sync" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.583524 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583530 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.583537 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="init" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583544 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="init" Nov 27 11:56:48 crc kubenswrapper[4796]: E1127 11:56:48.583556 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5200074-c987-416d-a17d-693d264d7ea8" containerName="keystone-bootstrap" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583562 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5200074-c987-416d-a17d-693d264d7ea8" containerName="keystone-bootstrap" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583782 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4924cd5e-791b-4312-b9bc-ad1167ec5fe4" containerName="neutron-db-sync" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583800 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.583813 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5200074-c987-416d-a17d-693d264d7ea8" containerName="keystone-bootstrap" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.584370 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.587125 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.587336 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.587358 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rc9hr" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.587372 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.588327 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.602101 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-svvgz"] Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.635174 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8whrf\" (UniqueName: \"kubernetes.io/projected/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-kube-api-access-8whrf\") pod \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.635251 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-combined-ca-bundle\") pod \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.635299 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-config\") pod \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\" (UID: \"4924cd5e-791b-4312-b9bc-ad1167ec5fe4\") " Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.641010 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-kube-api-access-8whrf" (OuterVolumeSpecName: "kube-api-access-8whrf") pod "4924cd5e-791b-4312-b9bc-ad1167ec5fe4" (UID: "4924cd5e-791b-4312-b9bc-ad1167ec5fe4"). InnerVolumeSpecName "kube-api-access-8whrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.660460 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4924cd5e-791b-4312-b9bc-ad1167ec5fe4" (UID: "4924cd5e-791b-4312-b9bc-ad1167ec5fe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.664792 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-config" (OuterVolumeSpecName: "config") pod "4924cd5e-791b-4312-b9bc-ad1167ec5fe4" (UID: "4924cd5e-791b-4312-b9bc-ad1167ec5fe4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.736999 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-scripts\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.737286 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-config-data\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.737432 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-credential-keys\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.737519 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-fernet-keys\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.737548 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl7g9\" (UniqueName: \"kubernetes.io/projected/6716271b-432f-4395-8d1a-3915979f3c33-kube-api-access-rl7g9\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.737759 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-combined-ca-bundle\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.738154 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8whrf\" (UniqueName: \"kubernetes.io/projected/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-kube-api-access-8whrf\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.738193 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.738208 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4924cd5e-791b-4312-b9bc-ad1167ec5fe4-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.814281 4796 scope.go:117] "RemoveContainer" containerID="4d78e6316eb3cb08efb46fa4783110ed5194c072a8b25766da32bea2abe11914" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.839769 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-fernet-keys\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.840117 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl7g9\" (UniqueName: \"kubernetes.io/projected/6716271b-432f-4395-8d1a-3915979f3c33-kube-api-access-rl7g9\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.840155 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-combined-ca-bundle\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.840280 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-scripts\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.840315 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-config-data\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.840350 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-credential-keys\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.856874 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-combined-ca-bundle\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.858065 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-scripts\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.858090 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-credential-keys\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.858105 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-fernet-keys\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.858194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-config-data\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.864184 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl7g9\" (UniqueName: \"kubernetes.io/projected/6716271b-432f-4395-8d1a-3915979f3c33-kube-api-access-rl7g9\") pod \"keystone-bootstrap-svvgz\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.872325 4796 scope.go:117] "RemoveContainer" containerID="34355bf4aac94f65b33f4030a233e523b4d7e7a9d1bb86f41f18d35a702e77bf" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.904591 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:48 crc kubenswrapper[4796]: I1127 11:56:48.910572 4796 scope.go:117] "RemoveContainer" containerID="3b64e2db5966a560814b7e9195f9f68f793526b34430de018558d8d04b2c193b" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.031611 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bbvsz" event={"ID":"4924cd5e-791b-4312-b9bc-ad1167ec5fe4","Type":"ContainerDied","Data":"22655bb8f0ef7bdd64bd0bff96072898d987a46bb6bc15b89094cf1e687fea42"} Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.031655 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22655bb8f0ef7bdd64bd0bff96072898d987a46bb6bc15b89094cf1e687fea42" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.031724 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bbvsz" Nov 27 11:56:49 crc kubenswrapper[4796]: E1127 11:56:49.075659 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-jfd4m" podUID="24d550fe-17da-4b68-8e62-b537d6c07314" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.174163 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-srjtx"] Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.190355 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.231432 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-srjtx"] Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.286676 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.319422 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55ffbfd4fd-lldkk"] Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.322929 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.329279 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-v6cdm" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.329443 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.329550 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.329646 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.343658 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55ffbfd4fd-lldkk"] Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.357895 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.358231 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.360255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7sf9\" (UniqueName: \"kubernetes.io/projected/4a56be28-0bcc-4635-ba4c-f96af33930a6-kube-api-access-h7sf9\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.360876 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-config\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.361456 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-svc\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.362827 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.375751 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469111 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-config\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469413 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469445 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469461 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7sf9\" (UniqueName: \"kubernetes.io/projected/4a56be28-0bcc-4635-ba4c-f96af33930a6-kube-api-access-h7sf9\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469482 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-config\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469501 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-svc\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469562 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469587 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-ovndb-tls-certs\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469626 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-combined-ca-bundle\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469670 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsvfd\" (UniqueName: \"kubernetes.io/projected/4eb23259-99e7-4823-8820-54eb455ae19f-kube-api-access-zsvfd\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.469686 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-httpd-config\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.470593 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.471080 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.471918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-config\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.471927 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.472534 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-svc\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.484480 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-svvgz"] Nov 27 11:56:49 crc kubenswrapper[4796]: W1127 11:56:49.492745 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6716271b_432f_4395_8d1a_3915979f3c33.slice/crio-e704590df8534d4dd380f5fa4f952798fb01743b7fb58ba522dc3a1e32091a82 WatchSource:0}: Error finding container e704590df8534d4dd380f5fa4f952798fb01743b7fb58ba522dc3a1e32091a82: Status 404 returned error can't find the container with id e704590df8534d4dd380f5fa4f952798fb01743b7fb58ba522dc3a1e32091a82 Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.504862 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7sf9\" (UniqueName: \"kubernetes.io/projected/4a56be28-0bcc-4635-ba4c-f96af33930a6-kube-api-access-h7sf9\") pod \"dnsmasq-dns-6b7b667979-srjtx\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.556189 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.571316 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-config\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.571438 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-ovndb-tls-certs\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.571460 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-combined-ca-bundle\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.571499 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsvfd\" (UniqueName: \"kubernetes.io/projected/4eb23259-99e7-4823-8820-54eb455ae19f-kube-api-access-zsvfd\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.571517 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-httpd-config\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.574658 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-config\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.574683 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-ovndb-tls-certs\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.578596 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-httpd-config\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.580393 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-combined-ca-bundle\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.595560 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5504fa9-43f7-442d-a025-245928896378" path="/var/lib/kubelet/pods/b5504fa9-43f7-442d-a025-245928896378/volumes" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.595905 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsvfd\" (UniqueName: \"kubernetes.io/projected/4eb23259-99e7-4823-8820-54eb455ae19f-kube-api-access-zsvfd\") pod \"neutron-55ffbfd4fd-lldkk\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.596244 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5200074-c987-416d-a17d-693d264d7ea8" path="/var/lib/kubelet/pods/e5200074-c987-416d-a17d-693d264d7ea8/volumes" Nov 27 11:56:49 crc kubenswrapper[4796]: I1127 11:56:49.649208 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.011597 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-ngtn7" podUID="b5504fa9-43f7-442d-a025-245928896378" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: i/o timeout" Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.119288 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dwbsz" event={"ID":"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3","Type":"ContainerStarted","Data":"7d773520869c92070138018c552043b76571c30f0d46205320a726857974f3cc"} Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.125688 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"453af780-fb5a-4048-b9fd-2ce7a497916c","Type":"ContainerStarted","Data":"480c745046ed9b4766f57c3c718ea1cedf1e084dff42b2e41053fc17d0ce460c"} Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.137581 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerStarted","Data":"295b69283f594d354574c7cb40ab077beea9846d5229d60a732e436dd705fa22"} Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.142723 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dwbsz" podStartSLOduration=3.272454581 podStartE2EDuration="29.142709229s" podCreationTimestamp="2025-11-27 11:56:21 +0000 UTC" firstStartedPulling="2025-11-27 11:56:22.549124758 +0000 UTC m=+1900.067443676" lastFinishedPulling="2025-11-27 11:56:48.419379406 +0000 UTC m=+1925.937698324" observedRunningTime="2025-11-27 11:56:50.138766554 +0000 UTC m=+1927.657085492" watchObservedRunningTime="2025-11-27 11:56:50.142709229 +0000 UTC m=+1927.661028147" Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.151809 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2960ce96-6701-446e-a800-6f5f9333a322","Type":"ContainerStarted","Data":"2b8ad5f98f8adf10585dacc5ef003217186af7733adf74230db95a1ec7b8aa5b"} Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.159499 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-svvgz" event={"ID":"6716271b-432f-4395-8d1a-3915979f3c33","Type":"ContainerStarted","Data":"2a542d127190aeed44c3c18627631a7b90870fa84e2b3f0cd41e93bc8514f27d"} Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.159546 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-svvgz" event={"ID":"6716271b-432f-4395-8d1a-3915979f3c33","Type":"ContainerStarted","Data":"e704590df8534d4dd380f5fa4f952798fb01743b7fb58ba522dc3a1e32091a82"} Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.184599 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-svvgz" podStartSLOduration=2.184584916 podStartE2EDuration="2.184584916s" podCreationTimestamp="2025-11-27 11:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:50.178320259 +0000 UTC m=+1927.696639187" watchObservedRunningTime="2025-11-27 11:56:50.184584916 +0000 UTC m=+1927.702903834" Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.278781 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-srjtx"] Nov 27 11:56:50 crc kubenswrapper[4796]: W1127 11:56:50.286450 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a56be28_0bcc_4635_ba4c_f96af33930a6.slice/crio-01b47c57a5b7fdf16a59aab60fd10b6e811cf3df10c93aa0497f37d0c76a9866 WatchSource:0}: Error finding container 01b47c57a5b7fdf16a59aab60fd10b6e811cf3df10c93aa0497f37d0c76a9866: Status 404 returned error can't find the container with id 01b47c57a5b7fdf16a59aab60fd10b6e811cf3df10c93aa0497f37d0c76a9866 Nov 27 11:56:50 crc kubenswrapper[4796]: I1127 11:56:50.507903 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55ffbfd4fd-lldkk"] Nov 27 11:56:50 crc kubenswrapper[4796]: W1127 11:56:50.543196 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eb23259_99e7_4823_8820_54eb455ae19f.slice/crio-75f5e8df808ed56133663fb2788267e1c020b80879e5d5c3222ca9f6428cacae WatchSource:0}: Error finding container 75f5e8df808ed56133663fb2788267e1c020b80879e5d5c3222ca9f6428cacae: Status 404 returned error can't find the container with id 75f5e8df808ed56133663fb2788267e1c020b80879e5d5c3222ca9f6428cacae Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.186378 4796 generic.go:334] "Generic (PLEG): container finished" podID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerID="73005fa7108240626f3d0bc59bc409209b418fb08e9218cc884a45a77963ed85" exitCode=0 Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.186814 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" event={"ID":"4a56be28-0bcc-4635-ba4c-f96af33930a6","Type":"ContainerDied","Data":"73005fa7108240626f3d0bc59bc409209b418fb08e9218cc884a45a77963ed85"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.186845 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" event={"ID":"4a56be28-0bcc-4635-ba4c-f96af33930a6","Type":"ContainerStarted","Data":"01b47c57a5b7fdf16a59aab60fd10b6e811cf3df10c93aa0497f37d0c76a9866"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.198035 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55ffbfd4fd-lldkk" event={"ID":"4eb23259-99e7-4823-8820-54eb455ae19f","Type":"ContainerStarted","Data":"77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.198088 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55ffbfd4fd-lldkk" event={"ID":"4eb23259-99e7-4823-8820-54eb455ae19f","Type":"ContainerStarted","Data":"dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.198101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55ffbfd4fd-lldkk" event={"ID":"4eb23259-99e7-4823-8820-54eb455ae19f","Type":"ContainerStarted","Data":"75f5e8df808ed56133663fb2788267e1c020b80879e5d5c3222ca9f6428cacae"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.199014 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.231495 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"453af780-fb5a-4048-b9fd-2ce7a497916c","Type":"ContainerStarted","Data":"e8fb4261492dd9c8fb3150ffa8de4241e497c1e9615dfe5b802ed25978cdb786"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.231539 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"453af780-fb5a-4048-b9fd-2ce7a497916c","Type":"ContainerStarted","Data":"8acd37e75f4453b24d2135a0ebba5b36b522d63e4fc64354e61a2026c4c1d889"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.242881 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2960ce96-6701-446e-a800-6f5f9333a322","Type":"ContainerStarted","Data":"378dff5c2638cd67b1de50ad9f7aad86648b95a18567b8218e9ed545b4a26020"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.242918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2960ce96-6701-446e-a800-6f5f9333a322","Type":"ContainerStarted","Data":"df3a543a9ac73d72ec17cbd51fe1220884beb423d558f79abd2ce1abf5c033e5"} Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.251171 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-55ffbfd4fd-lldkk" podStartSLOduration=2.251154379 podStartE2EDuration="2.251154379s" podCreationTimestamp="2025-11-27 11:56:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:51.231241828 +0000 UTC m=+1928.749560746" watchObservedRunningTime="2025-11-27 11:56:51.251154379 +0000 UTC m=+1928.769473297" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.266837 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.266815737 podStartE2EDuration="12.266815737s" podCreationTimestamp="2025-11-27 11:56:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:51.256293897 +0000 UTC m=+1928.774612835" watchObservedRunningTime="2025-11-27 11:56:51.266815737 +0000 UTC m=+1928.785134655" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.289116 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.289016089 podStartE2EDuration="11.289016089s" podCreationTimestamp="2025-11-27 11:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:51.283658496 +0000 UTC m=+1928.801977424" watchObservedRunningTime="2025-11-27 11:56:51.289016089 +0000 UTC m=+1928.807335007" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.672747 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69bc59f7f-g45f2"] Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.675413 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.677929 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.678240 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.686535 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69bc59f7f-g45f2"] Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.834101 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-internal-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.834151 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-httpd-config\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.834946 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-ovndb-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.834981 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-combined-ca-bundle\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.835027 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-config\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.835136 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-public-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.835356 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7pbj\" (UniqueName: \"kubernetes.io/projected/cdd9f25b-34db-477b-a6c6-11a5732cacdc-kube-api-access-g7pbj\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.937664 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-ovndb-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.937738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-combined-ca-bundle\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.937845 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-config\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.938053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-public-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.938893 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7pbj\" (UniqueName: \"kubernetes.io/projected/cdd9f25b-34db-477b-a6c6-11a5732cacdc-kube-api-access-g7pbj\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.938961 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-internal-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.938986 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-httpd-config\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.943647 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-combined-ca-bundle\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.943671 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-ovndb-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.944749 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-public-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.949754 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-internal-tls-certs\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.956766 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-config\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.962874 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cdd9f25b-34db-477b-a6c6-11a5732cacdc-httpd-config\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:51 crc kubenswrapper[4796]: I1127 11:56:51.966426 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7pbj\" (UniqueName: \"kubernetes.io/projected/cdd9f25b-34db-477b-a6c6-11a5732cacdc-kube-api-access-g7pbj\") pod \"neutron-69bc59f7f-g45f2\" (UID: \"cdd9f25b-34db-477b-a6c6-11a5732cacdc\") " pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:52 crc kubenswrapper[4796]: I1127 11:56:52.014187 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:52 crc kubenswrapper[4796]: I1127 11:56:52.265300 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" event={"ID":"4a56be28-0bcc-4635-ba4c-f96af33930a6","Type":"ContainerStarted","Data":"5609172066f2b057d1c47edcb79f73ce474e6c6381ee29b91daeb310f4815070"} Nov 27 11:56:52 crc kubenswrapper[4796]: I1127 11:56:52.265941 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:52 crc kubenswrapper[4796]: I1127 11:56:52.292447 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" podStartSLOduration=3.292417786 podStartE2EDuration="3.292417786s" podCreationTimestamp="2025-11-27 11:56:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:52.290816394 +0000 UTC m=+1929.809135322" watchObservedRunningTime="2025-11-27 11:56:52.292417786 +0000 UTC m=+1929.810736704" Nov 27 11:56:52 crc kubenswrapper[4796]: I1127 11:56:52.713488 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69bc59f7f-g45f2"] Nov 27 11:56:54 crc kubenswrapper[4796]: I1127 11:56:54.282867 4796 generic.go:334] "Generic (PLEG): container finished" podID="6716271b-432f-4395-8d1a-3915979f3c33" containerID="2a542d127190aeed44c3c18627631a7b90870fa84e2b3f0cd41e93bc8514f27d" exitCode=0 Nov 27 11:56:54 crc kubenswrapper[4796]: I1127 11:56:54.283254 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-svvgz" event={"ID":"6716271b-432f-4395-8d1a-3915979f3c33","Type":"ContainerDied","Data":"2a542d127190aeed44c3c18627631a7b90870fa84e2b3f0cd41e93bc8514f27d"} Nov 27 11:56:54 crc kubenswrapper[4796]: I1127 11:56:54.285404 4796 generic.go:334] "Generic (PLEG): container finished" podID="f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" containerID="7d773520869c92070138018c552043b76571c30f0d46205320a726857974f3cc" exitCode=0 Nov 27 11:56:54 crc kubenswrapper[4796]: I1127 11:56:54.285436 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dwbsz" event={"ID":"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3","Type":"ContainerDied","Data":"7d773520869c92070138018c552043b76571c30f0d46205320a726857974f3cc"} Nov 27 11:56:57 crc kubenswrapper[4796]: W1127 11:56:57.681796 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdd9f25b_34db_477b_a6c6_11a5732cacdc.slice/crio-fca12ee5f2d65830e4e09e0f8e4f7c8e213623f4c56830fa6b714742d99bd92d WatchSource:0}: Error finding container fca12ee5f2d65830e4e09e0f8e4f7c8e213623f4c56830fa6b714742d99bd92d: Status 404 returned error can't find the container with id fca12ee5f2d65830e4e09e0f8e4f7c8e213623f4c56830fa6b714742d99bd92d Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.051152 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.134146 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.149876 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-config-data\") pod \"6716271b-432f-4395-8d1a-3915979f3c33\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.149977 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-combined-ca-bundle\") pod \"6716271b-432f-4395-8d1a-3915979f3c33\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.150043 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-fernet-keys\") pod \"6716271b-432f-4395-8d1a-3915979f3c33\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.150096 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-scripts\") pod \"6716271b-432f-4395-8d1a-3915979f3c33\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.150150 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl7g9\" (UniqueName: \"kubernetes.io/projected/6716271b-432f-4395-8d1a-3915979f3c33-kube-api-access-rl7g9\") pod \"6716271b-432f-4395-8d1a-3915979f3c33\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.150181 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-credential-keys\") pod \"6716271b-432f-4395-8d1a-3915979f3c33\" (UID: \"6716271b-432f-4395-8d1a-3915979f3c33\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.155416 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6716271b-432f-4395-8d1a-3915979f3c33" (UID: "6716271b-432f-4395-8d1a-3915979f3c33"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.158918 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-scripts" (OuterVolumeSpecName: "scripts") pod "6716271b-432f-4395-8d1a-3915979f3c33" (UID: "6716271b-432f-4395-8d1a-3915979f3c33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.162212 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6716271b-432f-4395-8d1a-3915979f3c33-kube-api-access-rl7g9" (OuterVolumeSpecName: "kube-api-access-rl7g9") pod "6716271b-432f-4395-8d1a-3915979f3c33" (UID: "6716271b-432f-4395-8d1a-3915979f3c33"). InnerVolumeSpecName "kube-api-access-rl7g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.175891 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6716271b-432f-4395-8d1a-3915979f3c33" (UID: "6716271b-432f-4395-8d1a-3915979f3c33"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.226014 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6716271b-432f-4395-8d1a-3915979f3c33" (UID: "6716271b-432f-4395-8d1a-3915979f3c33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.227232 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-config-data" (OuterVolumeSpecName: "config-data") pod "6716271b-432f-4395-8d1a-3915979f3c33" (UID: "6716271b-432f-4395-8d1a-3915979f3c33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.251473 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-combined-ca-bundle\") pod \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.251617 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-db-sync-config-data\") pod \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.251661 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfkfq\" (UniqueName: \"kubernetes.io/projected/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-kube-api-access-sfkfq\") pod \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\" (UID: \"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3\") " Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.252077 4796 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.252093 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.252107 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.252118 4796 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.252129 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6716271b-432f-4395-8d1a-3915979f3c33-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.252139 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl7g9\" (UniqueName: \"kubernetes.io/projected/6716271b-432f-4395-8d1a-3915979f3c33-kube-api-access-rl7g9\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.255567 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-kube-api-access-sfkfq" (OuterVolumeSpecName: "kube-api-access-sfkfq") pod "f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" (UID: "f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3"). InnerVolumeSpecName "kube-api-access-sfkfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.258529 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" (UID: "f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.280859 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" (UID: "f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.342428 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-276z6" event={"ID":"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8","Type":"ContainerStarted","Data":"8570a2265dd638c975fe57479b98f4c4297282d9a1715f2d868a1ef50ab1210b"} Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.344650 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-svvgz" event={"ID":"6716271b-432f-4395-8d1a-3915979f3c33","Type":"ContainerDied","Data":"e704590df8534d4dd380f5fa4f952798fb01743b7fb58ba522dc3a1e32091a82"} Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.344720 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e704590df8534d4dd380f5fa4f952798fb01743b7fb58ba522dc3a1e32091a82" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.344669 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-svvgz" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.346291 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dwbsz" event={"ID":"f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3","Type":"ContainerDied","Data":"6f9fcfb451eac04360b1763d55bf189516b735156163ab9293162721cded688a"} Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.346336 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dwbsz" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.346338 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f9fcfb451eac04360b1763d55bf189516b735156163ab9293162721cded688a" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.347895 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69bc59f7f-g45f2" event={"ID":"cdd9f25b-34db-477b-a6c6-11a5732cacdc","Type":"ContainerStarted","Data":"7a899ae3edee1a9534718c5dfc289a5bf8663fb7f4d40d2e819e506b670377f9"} Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.347926 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69bc59f7f-g45f2" event={"ID":"cdd9f25b-34db-477b-a6c6-11a5732cacdc","Type":"ContainerStarted","Data":"fca12ee5f2d65830e4e09e0f8e4f7c8e213623f4c56830fa6b714742d99bd92d"} Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.349467 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerStarted","Data":"cb62072977045640d6230c54ad05fc56438fd69f2f79f3735f67ca9039ca6334"} Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.355504 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.355550 4796 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.355563 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfkfq\" (UniqueName: \"kubernetes.io/projected/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3-kube-api-access-sfkfq\") on node \"crc\" DevicePath \"\"" Nov 27 11:56:58 crc kubenswrapper[4796]: I1127 11:56:58.382416 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-276z6" podStartSLOduration=8.035358853 podStartE2EDuration="38.382394508s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="2025-11-27 11:56:22.314007685 +0000 UTC m=+1899.832326603" lastFinishedPulling="2025-11-27 11:56:52.66104334 +0000 UTC m=+1930.179362258" observedRunningTime="2025-11-27 11:56:58.375580667 +0000 UTC m=+1935.893899595" watchObservedRunningTime="2025-11-27 11:56:58.382394508 +0000 UTC m=+1935.900713426" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.178600 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-95c9f9f9b-zshfm"] Nov 27 11:56:59 crc kubenswrapper[4796]: E1127 11:56:59.179163 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" containerName="barbican-db-sync" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.179358 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" containerName="barbican-db-sync" Nov 27 11:56:59 crc kubenswrapper[4796]: E1127 11:56:59.179371 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6716271b-432f-4395-8d1a-3915979f3c33" containerName="keystone-bootstrap" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.179377 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6716271b-432f-4395-8d1a-3915979f3c33" containerName="keystone-bootstrap" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.179725 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6716271b-432f-4395-8d1a-3915979f3c33" containerName="keystone-bootstrap" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.179737 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" containerName="barbican-db-sync" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.185454 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.195364 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.195728 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.195902 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.196002 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rc9hr" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.196095 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.196198 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.212352 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-95c9f9f9b-zshfm"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277344 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-fernet-keys\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277401 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwk46\" (UniqueName: \"kubernetes.io/projected/0e86b961-a118-4ef7-a58a-a1a350c48783-kube-api-access-kwk46\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277497 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-internal-tls-certs\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277524 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-public-tls-certs\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277559 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-credential-keys\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277648 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-combined-ca-bundle\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277671 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-scripts\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.277693 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-config-data\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.359648 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69bc59f7f-g45f2" event={"ID":"cdd9f25b-34db-477b-a6c6-11a5732cacdc","Type":"ContainerStarted","Data":"72012df3f22e137e5f618d08c69002b271a41c5e38e9510a0d1058271a411e7c"} Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.360683 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.365883 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7968f8455-ntchl"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.367348 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.370864 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.378958 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-internal-tls-certs\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379425 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-public-tls-certs\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379504 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-credential-keys\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379670 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-combined-ca-bundle\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379695 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-scripts\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379715 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-config-data\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379791 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-fernet-keys\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.379808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwk46\" (UniqueName: \"kubernetes.io/projected/0e86b961-a118-4ef7-a58a-a1a350c48783-kube-api-access-kwk46\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.371859 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.371920 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9kkpt" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.380582 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7968f8455-ntchl"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.398748 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69bc59f7f-g45f2" podStartSLOduration=8.398732751 podStartE2EDuration="8.398732751s" podCreationTimestamp="2025-11-27 11:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:56:59.384941503 +0000 UTC m=+1936.903260421" watchObservedRunningTime="2025-11-27 11:56:59.398732751 +0000 UTC m=+1936.917051679" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.426385 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-combined-ca-bundle\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.426967 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-fernet-keys\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.427510 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-config-data\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.427737 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-scripts\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.428070 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-public-tls-certs\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.428460 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-internal-tls-certs\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.428847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwk46\" (UniqueName: \"kubernetes.io/projected/0e86b961-a118-4ef7-a58a-a1a350c48783-kube-api-access-kwk46\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.429207 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0e86b961-a118-4ef7-a58a-a1a350c48783-credential-keys\") pod \"keystone-95c9f9f9b-zshfm\" (UID: \"0e86b961-a118-4ef7-a58a-a1a350c48783\") " pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.487952 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176ad434-447c-4540-ac74-d80d1aeab2fd-logs\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.488025 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjxpr\" (UniqueName: \"kubernetes.io/projected/176ad434-447c-4540-ac74-d80d1aeab2fd-kube-api-access-tjxpr\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.488082 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-combined-ca-bundle\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.488153 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-config-data\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.488290 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-config-data-custom\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.517814 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.566599 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-79c6986754-nzz2p"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.568107 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.568166 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.579211 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.591772 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-config-data-custom\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.591855 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176ad434-447c-4540-ac74-d80d1aeab2fd-logs\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.592145 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjxpr\" (UniqueName: \"kubernetes.io/projected/176ad434-447c-4540-ac74-d80d1aeab2fd-kube-api-access-tjxpr\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.592173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-combined-ca-bundle\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.592209 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-config-data\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.603408 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176ad434-447c-4540-ac74-d80d1aeab2fd-logs\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.607396 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-config-data-custom\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.614110 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-config-data\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.625010 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79c6986754-nzz2p"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.629144 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176ad434-447c-4540-ac74-d80d1aeab2fd-combined-ca-bundle\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.645858 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjxpr\" (UniqueName: \"kubernetes.io/projected/176ad434-447c-4540-ac74-d80d1aeab2fd-kube-api-access-tjxpr\") pod \"barbican-worker-7968f8455-ntchl\" (UID: \"176ad434-447c-4540-ac74-d80d1aeab2fd\") " pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.663066 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-srjtx"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.693342 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zwrc6"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.693659 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-combined-ca-bundle\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.693760 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-config-data-custom\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.693883 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrhmk\" (UniqueName: \"kubernetes.io/projected/c0327df8-c909-4f47-bff7-7519ded4ea2a-kube-api-access-nrhmk\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.693927 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-config-data\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.693959 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0327df8-c909-4f47-bff7-7519ded4ea2a-logs\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.695189 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.704698 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zwrc6"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.772393 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5dcd9f4978-tl45d"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.776854 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.778866 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.790695 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dcd9f4978-tl45d"] Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798283 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-config-data-custom\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798446 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798487 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrhmk\" (UniqueName: \"kubernetes.io/projected/c0327df8-c909-4f47-bff7-7519ded4ea2a-kube-api-access-nrhmk\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798513 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798537 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798557 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-config-data\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798573 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-config\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798598 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0327df8-c909-4f47-bff7-7519ded4ea2a-logs\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798631 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fccgc\" (UniqueName: \"kubernetes.io/projected/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-kube-api-access-fccgc\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798659 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.798687 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-combined-ca-bundle\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.803649 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0327df8-c909-4f47-bff7-7519ded4ea2a-logs\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.805514 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-config-data-custom\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.807107 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-config-data\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.822801 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7968f8455-ntchl" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.822950 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0327df8-c909-4f47-bff7-7519ded4ea2a-combined-ca-bundle\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.831152 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrhmk\" (UniqueName: \"kubernetes.io/projected/c0327df8-c909-4f47-bff7-7519ded4ea2a-kube-api-access-nrhmk\") pod \"barbican-keystone-listener-79c6986754-nzz2p\" (UID: \"c0327df8-c909-4f47-bff7-7519ded4ea2a\") " pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.902399 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data-custom\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903108 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-config\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-logs\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903381 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-combined-ca-bundle\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903436 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fccgc\" (UniqueName: \"kubernetes.io/projected/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-kube-api-access-fccgc\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903489 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903699 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903750 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v27vq\" (UniqueName: \"kubernetes.io/projected/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-kube-api-access-v27vq\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903813 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903835 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.903863 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.904675 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-config\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.907103 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.908004 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.908631 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.908765 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.928722 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fccgc\" (UniqueName: \"kubernetes.io/projected/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-kube-api-access-fccgc\") pod \"dnsmasq-dns-848cf88cfc-zwrc6\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:56:59 crc kubenswrapper[4796]: I1127 11:56:59.938209 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-95c9f9f9b-zshfm"] Nov 27 11:56:59 crc kubenswrapper[4796]: W1127 11:56:59.951343 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e86b961_a118_4ef7_a58a_a1a350c48783.slice/crio-9d62280c93c64c9fce9ad59827b7c5a2336d164475f9dca1c5dafb9f12dd3158 WatchSource:0}: Error finding container 9d62280c93c64c9fce9ad59827b7c5a2336d164475f9dca1c5dafb9f12dd3158: Status 404 returned error can't find the container with id 9d62280c93c64c9fce9ad59827b7c5a2336d164475f9dca1c5dafb9f12dd3158 Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.006509 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.006545 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data-custom\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.006564 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-logs\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.006590 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-combined-ca-bundle\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.006721 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v27vq\" (UniqueName: \"kubernetes.io/projected/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-kube-api-access-v27vq\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.012488 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-logs\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.015165 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.016541 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-combined-ca-bundle\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.017882 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data-custom\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.020063 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.024038 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v27vq\" (UniqueName: \"kubernetes.io/projected/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-kube-api-access-v27vq\") pod \"barbican-api-5dcd9f4978-tl45d\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.060878 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.097850 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.363577 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7968f8455-ntchl"] Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.388517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-95c9f9f9b-zshfm" event={"ID":"0e86b961-a118-4ef7-a58a-a1a350c48783","Type":"ContainerStarted","Data":"9d62280c93c64c9fce9ad59827b7c5a2336d164475f9dca1c5dafb9f12dd3158"} Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.388708 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerName="dnsmasq-dns" containerID="cri-o://5609172066f2b057d1c47edcb79f73ce474e6c6381ee29b91daeb310f4815070" gracePeriod=10 Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.515968 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79c6986754-nzz2p"] Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.611354 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.611398 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.621584 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.622566 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.644602 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.659999 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.690544 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.714380 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.744518 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dcd9f4978-tl45d"] Nov 27 11:57:00 crc kubenswrapper[4796]: I1127 11:57:00.789338 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zwrc6"] Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.396780 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" event={"ID":"1a852724-a4bb-4ed6-aa8e-da99f2d31de2","Type":"ContainerStarted","Data":"7659b6edf86ced3a07c3b826a82c5a234fa81610f97652da66d43e603802a7fc"} Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.398528 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dcd9f4978-tl45d" event={"ID":"91118dbe-a3ca-41b6-ab86-a24ebd7b5321","Type":"ContainerStarted","Data":"129d7511e21e2e3d15660f4d0d24573627d437820fe521be88fcfdda9392addd"} Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.399567 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" event={"ID":"c0327df8-c909-4f47-bff7-7519ded4ea2a","Type":"ContainerStarted","Data":"d6eb281adaa06d7a22b08d5caf75f59d48ae7f7eb79fdb265d7b86f3817b3500"} Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.400610 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7968f8455-ntchl" event={"ID":"176ad434-447c-4540-ac74-d80d1aeab2fd","Type":"ContainerStarted","Data":"b39d319a0ebf538e570198d3c3687acac162a00837b27fd00234f22b41cd822b"} Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.402758 4796 generic.go:334] "Generic (PLEG): container finished" podID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerID="5609172066f2b057d1c47edcb79f73ce474e6c6381ee29b91daeb310f4815070" exitCode=0 Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.402780 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" event={"ID":"4a56be28-0bcc-4635-ba4c-f96af33930a6","Type":"ContainerDied","Data":"5609172066f2b057d1c47edcb79f73ce474e6c6381ee29b91daeb310f4815070"} Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.403414 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.403474 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.403743 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.403785 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 11:57:01 crc kubenswrapper[4796]: I1127 11:57:01.573377 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:57:01 crc kubenswrapper[4796]: E1127 11:57:01.573804 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.175901 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.271291 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7sf9\" (UniqueName: \"kubernetes.io/projected/4a56be28-0bcc-4635-ba4c-f96af33930a6-kube-api-access-h7sf9\") pod \"4a56be28-0bcc-4635-ba4c-f96af33930a6\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.272377 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-svc\") pod \"4a56be28-0bcc-4635-ba4c-f96af33930a6\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.272518 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-sb\") pod \"4a56be28-0bcc-4635-ba4c-f96af33930a6\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.272632 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-nb\") pod \"4a56be28-0bcc-4635-ba4c-f96af33930a6\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.273202 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-config\") pod \"4a56be28-0bcc-4635-ba4c-f96af33930a6\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.273416 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-swift-storage-0\") pod \"4a56be28-0bcc-4635-ba4c-f96af33930a6\" (UID: \"4a56be28-0bcc-4635-ba4c-f96af33930a6\") " Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.277748 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a56be28-0bcc-4635-ba4c-f96af33930a6-kube-api-access-h7sf9" (OuterVolumeSpecName: "kube-api-access-h7sf9") pod "4a56be28-0bcc-4635-ba4c-f96af33930a6" (UID: "4a56be28-0bcc-4635-ba4c-f96af33930a6"). InnerVolumeSpecName "kube-api-access-h7sf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.381545 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7sf9\" (UniqueName: \"kubernetes.io/projected/4a56be28-0bcc-4635-ba4c-f96af33930a6-kube-api-access-h7sf9\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.420365 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" event={"ID":"4a56be28-0bcc-4635-ba4c-f96af33930a6","Type":"ContainerDied","Data":"01b47c57a5b7fdf16a59aab60fd10b6e811cf3df10c93aa0497f37d0c76a9866"} Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.420419 4796 scope.go:117] "RemoveContainer" containerID="5609172066f2b057d1c47edcb79f73ce474e6c6381ee29b91daeb310f4815070" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.420565 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-srjtx" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.424651 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-config" (OuterVolumeSpecName: "config") pod "4a56be28-0bcc-4635-ba4c-f96af33930a6" (UID: "4a56be28-0bcc-4635-ba4c-f96af33930a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.430722 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4a56be28-0bcc-4635-ba4c-f96af33930a6" (UID: "4a56be28-0bcc-4635-ba4c-f96af33930a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.430913 4796 generic.go:334] "Generic (PLEG): container finished" podID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerID="e6bf0c7fc2f34a7e1c3d14584a6a358efcd59b6f258bbddeb005e0fe737f0db2" exitCode=0 Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.430996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" event={"ID":"1a852724-a4bb-4ed6-aa8e-da99f2d31de2","Type":"ContainerDied","Data":"e6bf0c7fc2f34a7e1c3d14584a6a358efcd59b6f258bbddeb005e0fe737f0db2"} Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.439850 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4a56be28-0bcc-4635-ba4c-f96af33930a6" (UID: "4a56be28-0bcc-4635-ba4c-f96af33930a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.444849 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4a56be28-0bcc-4635-ba4c-f96af33930a6" (UID: "4a56be28-0bcc-4635-ba4c-f96af33930a6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.466193 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4a56be28-0bcc-4635-ba4c-f96af33930a6" (UID: "4a56be28-0bcc-4635-ba4c-f96af33930a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.466882 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dcd9f4978-tl45d" event={"ID":"91118dbe-a3ca-41b6-ab86-a24ebd7b5321","Type":"ContainerStarted","Data":"5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c"} Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.466931 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dcd9f4978-tl45d" event={"ID":"91118dbe-a3ca-41b6-ab86-a24ebd7b5321","Type":"ContainerStarted","Data":"8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a"} Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.468144 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.468175 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.485846 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-95c9f9f9b-zshfm" event={"ID":"0e86b961-a118-4ef7-a58a-a1a350c48783","Type":"ContainerStarted","Data":"9e6a9de662360019bef6e660cf57a2d3d84339fec7dcce84a3be698161c2f418"} Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.485901 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.496814 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.496852 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.496864 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.496876 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.496902 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4a56be28-0bcc-4635-ba4c-f96af33930a6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.509431 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5dcd9f4978-tl45d" podStartSLOduration=3.509407805 podStartE2EDuration="3.509407805s" podCreationTimestamp="2025-11-27 11:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:02.499357047 +0000 UTC m=+1940.017675965" watchObservedRunningTime="2025-11-27 11:57:02.509407805 +0000 UTC m=+1940.027726723" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.573426 4796 scope.go:117] "RemoveContainer" containerID="73005fa7108240626f3d0bc59bc409209b418fb08e9218cc884a45a77963ed85" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.884758 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-95c9f9f9b-zshfm" podStartSLOduration=3.884739098 podStartE2EDuration="3.884739098s" podCreationTimestamp="2025-11-27 11:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:02.528245177 +0000 UTC m=+1940.046564105" watchObservedRunningTime="2025-11-27 11:57:02.884739098 +0000 UTC m=+1940.403058006" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.886627 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-srjtx"] Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.894478 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-srjtx"] Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.904397 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-655d9778b6-79p2x"] Nov 27 11:57:02 crc kubenswrapper[4796]: E1127 11:57:02.904747 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerName="dnsmasq-dns" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.904760 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerName="dnsmasq-dns" Nov 27 11:57:02 crc kubenswrapper[4796]: E1127 11:57:02.904782 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerName="init" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.904790 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerName="init" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.904992 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" containerName="dnsmasq-dns" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.906044 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.908582 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.908802 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 27 11:57:02 crc kubenswrapper[4796]: I1127 11:57:02.929152 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-655d9778b6-79p2x"] Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032480 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-public-tls-certs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032541 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c71cc6f0-df01-4c3d-9962-29d71307f38a-logs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032577 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-internal-tls-certs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032626 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-combined-ca-bundle\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032649 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-config-data\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032743 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-config-data-custom\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.032776 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn9cr\" (UniqueName: \"kubernetes.io/projected/c71cc6f0-df01-4c3d-9962-29d71307f38a-kube-api-access-tn9cr\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.134771 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-config-data-custom\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.134840 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn9cr\" (UniqueName: \"kubernetes.io/projected/c71cc6f0-df01-4c3d-9962-29d71307f38a-kube-api-access-tn9cr\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.134914 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-public-tls-certs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.134936 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c71cc6f0-df01-4c3d-9962-29d71307f38a-logs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.136153 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-internal-tls-certs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.136220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-combined-ca-bundle\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.136245 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-config-data\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.136505 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c71cc6f0-df01-4c3d-9962-29d71307f38a-logs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.142202 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-config-data\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.143365 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-config-data-custom\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.146941 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-internal-tls-certs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.151526 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-public-tls-certs\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.153037 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn9cr\" (UniqueName: \"kubernetes.io/projected/c71cc6f0-df01-4c3d-9962-29d71307f38a-kube-api-access-tn9cr\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.167898 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c71cc6f0-df01-4c3d-9962-29d71307f38a-combined-ca-bundle\") pod \"barbican-api-655d9778b6-79p2x\" (UID: \"c71cc6f0-df01-4c3d-9962-29d71307f38a\") " pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.301868 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.494030 4796 generic.go:334] "Generic (PLEG): container finished" podID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" containerID="8570a2265dd638c975fe57479b98f4c4297282d9a1715f2d868a1ef50ab1210b" exitCode=0 Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.494083 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-276z6" event={"ID":"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8","Type":"ContainerDied","Data":"8570a2265dd638c975fe57479b98f4c4297282d9a1715f2d868a1ef50ab1210b"} Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.501598 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-km9nk" event={"ID":"d920f619-249d-4266-ae4b-e13917ae78dc","Type":"ContainerStarted","Data":"7dc64cda3633c4e68703272a8f99797cb28b9dd31d7ef53257e67b9e59902190"} Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.505580 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfd4m" event={"ID":"24d550fe-17da-4b68-8e62-b537d6c07314","Type":"ContainerStarted","Data":"33445dd4e42e13ddd1b17ea17ed7681779e513f0033ca2caa47a57e772114b46"} Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.509123 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" event={"ID":"1a852724-a4bb-4ed6-aa8e-da99f2d31de2","Type":"ContainerStarted","Data":"20bfd9fd2b3fc49622ee7161c967ac41cf3b51caedaf10afa53ab77b17dd9043"} Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.509910 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.541000 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-jfd4m" podStartSLOduration=3.732560338 podStartE2EDuration="43.540979904s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="2025-11-27 11:56:22.313731988 +0000 UTC m=+1899.832050906" lastFinishedPulling="2025-11-27 11:57:02.122151554 +0000 UTC m=+1939.640470472" observedRunningTime="2025-11-27 11:57:03.535675192 +0000 UTC m=+1941.053994120" watchObservedRunningTime="2025-11-27 11:57:03.540979904 +0000 UTC m=+1941.059298822" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.586477 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" podStartSLOduration=4.586461967 podStartE2EDuration="4.586461967s" podCreationTimestamp="2025-11-27 11:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:03.571916859 +0000 UTC m=+1941.090235767" watchObservedRunningTime="2025-11-27 11:57:03.586461967 +0000 UTC m=+1941.104780885" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.588768 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-km9nk" podStartSLOduration=3.357251756 podStartE2EDuration="43.588755318s" podCreationTimestamp="2025-11-27 11:56:20 +0000 UTC" firstStartedPulling="2025-11-27 11:56:22.015243576 +0000 UTC m=+1899.533562494" lastFinishedPulling="2025-11-27 11:57:02.246747128 +0000 UTC m=+1939.765066056" observedRunningTime="2025-11-27 11:57:03.552855971 +0000 UTC m=+1941.071174899" watchObservedRunningTime="2025-11-27 11:57:03.588755318 +0000 UTC m=+1941.107074236" Nov 27 11:57:03 crc kubenswrapper[4796]: I1127 11:57:03.604826 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a56be28-0bcc-4635-ba4c-f96af33930a6" path="/var/lib/kubelet/pods/4a56be28-0bcc-4635-ba4c-f96af33930a6/volumes" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.065461 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.065572 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.237166 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.461005 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.461125 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.602983 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 11:57:04 crc kubenswrapper[4796]: I1127 11:57:04.981995 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-655d9778b6-79p2x"] Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.562003 4796 generic.go:334] "Generic (PLEG): container finished" podID="d920f619-249d-4266-ae4b-e13917ae78dc" containerID="7dc64cda3633c4e68703272a8f99797cb28b9dd31d7ef53257e67b9e59902190" exitCode=0 Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.562118 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-km9nk" event={"ID":"d920f619-249d-4266-ae4b-e13917ae78dc","Type":"ContainerDied","Data":"7dc64cda3633c4e68703272a8f99797cb28b9dd31d7ef53257e67b9e59902190"} Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.846430 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-276z6" Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.944440 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-config-data\") pod \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.944514 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-logs\") pod \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.944539 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8sv9\" (UniqueName: \"kubernetes.io/projected/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-kube-api-access-f8sv9\") pod \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.944632 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-combined-ca-bundle\") pod \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.944787 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-scripts\") pod \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\" (UID: \"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8\") " Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.944839 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-logs" (OuterVolumeSpecName: "logs") pod "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" (UID: "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.945184 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.950963 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-scripts" (OuterVolumeSpecName: "scripts") pod "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" (UID: "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.951083 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-kube-api-access-f8sv9" (OuterVolumeSpecName: "kube-api-access-f8sv9") pod "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" (UID: "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8"). InnerVolumeSpecName "kube-api-access-f8sv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.973176 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-config-data" (OuterVolumeSpecName: "config-data") pod "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" (UID: "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:07 crc kubenswrapper[4796]: I1127 11:57:07.983982 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" (UID: "23ffa4b0-e22e-4714-9b05-96bd3fffbfa8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.047405 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8sv9\" (UniqueName: \"kubernetes.io/projected/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-kube-api-access-f8sv9\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.047439 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.047448 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.047456 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:08 crc kubenswrapper[4796]: W1127 11:57:08.403973 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc71cc6f0_df01_4c3d_9962_29d71307f38a.slice/crio-3f3de2edf6184c9cfcf2c95c98398a7c65b5f505d829dfe89f907432ace73e0b WatchSource:0}: Error finding container 3f3de2edf6184c9cfcf2c95c98398a7c65b5f505d829dfe89f907432ace73e0b: Status 404 returned error can't find the container with id 3f3de2edf6184c9cfcf2c95c98398a7c65b5f505d829dfe89f907432ace73e0b Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.571443 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-276z6" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.571439 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-276z6" event={"ID":"23ffa4b0-e22e-4714-9b05-96bd3fffbfa8","Type":"ContainerDied","Data":"b049de2f72cdebded9e4c5f460de98593c5ca9e5dc60f8abc8c18aebf8dd583d"} Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.571576 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b049de2f72cdebded9e4c5f460de98593c5ca9e5dc60f8abc8c18aebf8dd583d" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.573289 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7968f8455-ntchl" event={"ID":"176ad434-447c-4540-ac74-d80d1aeab2fd","Type":"ContainerStarted","Data":"b447a8973e7a50d72c581d847cfe03753de12c78ea13b38e7115723fbbe4fe53"} Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.575516 4796 generic.go:334] "Generic (PLEG): container finished" podID="24d550fe-17da-4b68-8e62-b537d6c07314" containerID="33445dd4e42e13ddd1b17ea17ed7681779e513f0033ca2caa47a57e772114b46" exitCode=0 Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.575588 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfd4m" event={"ID":"24d550fe-17da-4b68-8e62-b537d6c07314","Type":"ContainerDied","Data":"33445dd4e42e13ddd1b17ea17ed7681779e513f0033ca2caa47a57e772114b46"} Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.577348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-655d9778b6-79p2x" event={"ID":"c71cc6f0-df01-4c3d-9962-29d71307f38a","Type":"ContainerStarted","Data":"3f3de2edf6184c9cfcf2c95c98398a7c65b5f505d829dfe89f907432ace73e0b"} Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.938392 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d4b54c868-49fn9"] Nov 27 11:57:08 crc kubenswrapper[4796]: E1127 11:57:08.939015 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" containerName="placement-db-sync" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.939035 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" containerName="placement-db-sync" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.939250 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" containerName="placement-db-sync" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.940114 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.942759 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.943057 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.943196 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.943471 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mj92m" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.943732 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963403 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-combined-ca-bundle\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963449 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-scripts\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963507 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-public-tls-certs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963555 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-config-data\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963602 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-internal-tls-certs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963625 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9233c70-4c77-41e8-831b-682307ae7dd7-logs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.963652 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtg6t\" (UniqueName: \"kubernetes.io/projected/c9233c70-4c77-41e8-831b-682307ae7dd7-kube-api-access-rtg6t\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:08 crc kubenswrapper[4796]: I1127 11:57:08.974521 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d4b54c868-49fn9"] Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.006687 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-km9nk" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.064832 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnbv4\" (UniqueName: \"kubernetes.io/projected/d920f619-249d-4266-ae4b-e13917ae78dc-kube-api-access-xnbv4\") pod \"d920f619-249d-4266-ae4b-e13917ae78dc\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.064948 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-config-data\") pod \"d920f619-249d-4266-ae4b-e13917ae78dc\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065037 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-combined-ca-bundle\") pod \"d920f619-249d-4266-ae4b-e13917ae78dc\" (UID: \"d920f619-249d-4266-ae4b-e13917ae78dc\") " Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065362 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-public-tls-certs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065439 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-config-data\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065496 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-internal-tls-certs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065522 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9233c70-4c77-41e8-831b-682307ae7dd7-logs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065552 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtg6t\" (UniqueName: \"kubernetes.io/projected/c9233c70-4c77-41e8-831b-682307ae7dd7-kube-api-access-rtg6t\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065581 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-combined-ca-bundle\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.065601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-scripts\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.067247 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9233c70-4c77-41e8-831b-682307ae7dd7-logs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.069106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-scripts\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.072494 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-public-tls-certs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.073129 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-combined-ca-bundle\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.074791 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-config-data\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.077509 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9233c70-4c77-41e8-831b-682307ae7dd7-internal-tls-certs\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.081053 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d920f619-249d-4266-ae4b-e13917ae78dc-kube-api-access-xnbv4" (OuterVolumeSpecName: "kube-api-access-xnbv4") pod "d920f619-249d-4266-ae4b-e13917ae78dc" (UID: "d920f619-249d-4266-ae4b-e13917ae78dc"). InnerVolumeSpecName "kube-api-access-xnbv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.086088 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtg6t\" (UniqueName: \"kubernetes.io/projected/c9233c70-4c77-41e8-831b-682307ae7dd7-kube-api-access-rtg6t\") pod \"placement-7d4b54c868-49fn9\" (UID: \"c9233c70-4c77-41e8-831b-682307ae7dd7\") " pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.104951 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d920f619-249d-4266-ae4b-e13917ae78dc" (UID: "d920f619-249d-4266-ae4b-e13917ae78dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.136473 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.168974 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.169001 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnbv4\" (UniqueName: \"kubernetes.io/projected/d920f619-249d-4266-ae4b-e13917ae78dc-kube-api-access-xnbv4\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.193430 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-config-data" (OuterVolumeSpecName: "config-data") pod "d920f619-249d-4266-ae4b-e13917ae78dc" (UID: "d920f619-249d-4266-ae4b-e13917ae78dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.296408 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d920f619-249d-4266-ae4b-e13917ae78dc-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:09 crc kubenswrapper[4796]: E1127 11:57:09.330940 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.601689 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" event={"ID":"c0327df8-c909-4f47-bff7-7519ded4ea2a","Type":"ContainerStarted","Data":"b6c40d26da6eea08be8adbf36ddf941c4e3478604309e67e399470a0a500c227"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.601772 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" event={"ID":"c0327df8-c909-4f47-bff7-7519ded4ea2a","Type":"ContainerStarted","Data":"8f7197da4a45774806142fab6d1c1eceed3288f0118565cb14c4685d5819ed61"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.605610 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-km9nk" event={"ID":"d920f619-249d-4266-ae4b-e13917ae78dc","Type":"ContainerDied","Data":"bbadb8bb7ea5a4e48d9e5ac79e58f16d5c4da96e50285d010ceca03e437d4382"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.605647 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbadb8bb7ea5a4e48d9e5ac79e58f16d5c4da96e50285d010ceca03e437d4382" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.605722 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-km9nk" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.611082 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7968f8455-ntchl" event={"ID":"176ad434-447c-4540-ac74-d80d1aeab2fd","Type":"ContainerStarted","Data":"28731ca6b4a81248d7f9dda1eefc09021e8b523e77522bb23f70da1f689c4673"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.621383 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerStarted","Data":"3c186646c11391054c8c86afc01f80accaeb1dccd4d12fdea3aa7e204727af4a"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.621411 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="ceilometer-notification-agent" containerID="cri-o://295b69283f594d354574c7cb40ab077beea9846d5229d60a732e436dd705fa22" gracePeriod=30 Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.621584 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.621626 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="proxy-httpd" containerID="cri-o://3c186646c11391054c8c86afc01f80accaeb1dccd4d12fdea3aa7e204727af4a" gracePeriod=30 Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.621636 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="sg-core" containerID="cri-o://cb62072977045640d6230c54ad05fc56438fd69f2f79f3735f67ca9039ca6334" gracePeriod=30 Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.632203 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-79c6986754-nzz2p" podStartSLOduration=2.290064406 podStartE2EDuration="10.632183988s" podCreationTimestamp="2025-11-27 11:56:59 +0000 UTC" firstStartedPulling="2025-11-27 11:57:00.537394928 +0000 UTC m=+1938.055713846" lastFinishedPulling="2025-11-27 11:57:08.87951451 +0000 UTC m=+1946.397833428" observedRunningTime="2025-11-27 11:57:09.623004124 +0000 UTC m=+1947.141323042" watchObservedRunningTime="2025-11-27 11:57:09.632183988 +0000 UTC m=+1947.150502896" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.644474 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-655d9778b6-79p2x" event={"ID":"c71cc6f0-df01-4c3d-9962-29d71307f38a","Type":"ContainerStarted","Data":"22fbe26d0d718afbe29ffe5868ff0dc0bb8cbf43650d334d03cd96942a53a883"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.644513 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.644523 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-655d9778b6-79p2x" event={"ID":"c71cc6f0-df01-4c3d-9962-29d71307f38a","Type":"ContainerStarted","Data":"42c57918f8637aa89f054c7e645afd9c7bfeea9f5d06884016fdf93e93486848"} Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.644543 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.682815 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7968f8455-ntchl" podStartSLOduration=6.649470321 podStartE2EDuration="10.682794188s" podCreationTimestamp="2025-11-27 11:56:59 +0000 UTC" firstStartedPulling="2025-11-27 11:57:00.383691447 +0000 UTC m=+1937.902010365" lastFinishedPulling="2025-11-27 11:57:04.417015314 +0000 UTC m=+1941.935334232" observedRunningTime="2025-11-27 11:57:09.672138734 +0000 UTC m=+1947.190457662" watchObservedRunningTime="2025-11-27 11:57:09.682794188 +0000 UTC m=+1947.201113106" Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.683014 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d4b54c868-49fn9"] Nov 27 11:57:09 crc kubenswrapper[4796]: I1127 11:57:09.735425 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-655d9778b6-79p2x" podStartSLOduration=7.735408202 podStartE2EDuration="7.735408202s" podCreationTimestamp="2025-11-27 11:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:09.720621838 +0000 UTC m=+1947.238940756" watchObservedRunningTime="2025-11-27 11:57:09.735408202 +0000 UTC m=+1947.253727120" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.062422 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.141594 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-mkwn4"] Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.141828 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerName="dnsmasq-dns" containerID="cri-o://eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b" gracePeriod=10 Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.227444 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.317043 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-config-data\") pod \"24d550fe-17da-4b68-8e62-b537d6c07314\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.317115 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-scripts\") pod \"24d550fe-17da-4b68-8e62-b537d6c07314\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.317135 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24d550fe-17da-4b68-8e62-b537d6c07314-etc-machine-id\") pod \"24d550fe-17da-4b68-8e62-b537d6c07314\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.317418 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84tsj\" (UniqueName: \"kubernetes.io/projected/24d550fe-17da-4b68-8e62-b537d6c07314-kube-api-access-84tsj\") pod \"24d550fe-17da-4b68-8e62-b537d6c07314\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.317506 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-combined-ca-bundle\") pod \"24d550fe-17da-4b68-8e62-b537d6c07314\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.317578 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-db-sync-config-data\") pod \"24d550fe-17da-4b68-8e62-b537d6c07314\" (UID: \"24d550fe-17da-4b68-8e62-b537d6c07314\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.319894 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24d550fe-17da-4b68-8e62-b537d6c07314-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "24d550fe-17da-4b68-8e62-b537d6c07314" (UID: "24d550fe-17da-4b68-8e62-b537d6c07314"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.323321 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-scripts" (OuterVolumeSpecName: "scripts") pod "24d550fe-17da-4b68-8e62-b537d6c07314" (UID: "24d550fe-17da-4b68-8e62-b537d6c07314"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.323665 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "24d550fe-17da-4b68-8e62-b537d6c07314" (UID: "24d550fe-17da-4b68-8e62-b537d6c07314"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.323682 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d550fe-17da-4b68-8e62-b537d6c07314-kube-api-access-84tsj" (OuterVolumeSpecName: "kube-api-access-84tsj") pod "24d550fe-17da-4b68-8e62-b537d6c07314" (UID: "24d550fe-17da-4b68-8e62-b537d6c07314"). InnerVolumeSpecName "kube-api-access-84tsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.381949 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24d550fe-17da-4b68-8e62-b537d6c07314" (UID: "24d550fe-17da-4b68-8e62-b537d6c07314"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.419623 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.419652 4796 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24d550fe-17da-4b68-8e62-b537d6c07314-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.419669 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84tsj\" (UniqueName: \"kubernetes.io/projected/24d550fe-17da-4b68-8e62-b537d6c07314-kube-api-access-84tsj\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.419680 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.419691 4796 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.427786 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-config-data" (OuterVolumeSpecName: "config-data") pod "24d550fe-17da-4b68-8e62-b537d6c07314" (UID: "24d550fe-17da-4b68-8e62-b537d6c07314"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.521516 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d550fe-17da-4b68-8e62-b537d6c07314-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.657364 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.662168 4796 generic.go:334] "Generic (PLEG): container finished" podID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerID="eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b" exitCode=0 Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.662224 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" event={"ID":"680cc550-1bfc-4a40-97e5-72067cb9efb7","Type":"ContainerDied","Data":"eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.662248 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" event={"ID":"680cc550-1bfc-4a40-97e5-72067cb9efb7","Type":"ContainerDied","Data":"697e421b63756cae1b525b76c4b1f66075f24b7d15a817f550581400aff96eec"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.662292 4796 scope.go:117] "RemoveContainer" containerID="eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.676503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jfd4m" event={"ID":"24d550fe-17da-4b68-8e62-b537d6c07314","Type":"ContainerDied","Data":"f99347f3a40580cf78a3f7bc7d85f436fe32bf185d01c56760bc7e5111a264a2"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.676562 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99347f3a40580cf78a3f7bc7d85f436fe32bf185d01c56760bc7e5111a264a2" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.676725 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jfd4m" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.694416 4796 scope.go:117] "RemoveContainer" containerID="2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.706927 4796 generic.go:334] "Generic (PLEG): container finished" podID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerID="3c186646c11391054c8c86afc01f80accaeb1dccd4d12fdea3aa7e204727af4a" exitCode=0 Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.707124 4796 generic.go:334] "Generic (PLEG): container finished" podID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerID="cb62072977045640d6230c54ad05fc56438fd69f2f79f3735f67ca9039ca6334" exitCode=2 Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.707218 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerDied","Data":"3c186646c11391054c8c86afc01f80accaeb1dccd4d12fdea3aa7e204727af4a"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.707329 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerDied","Data":"cb62072977045640d6230c54ad05fc56438fd69f2f79f3735f67ca9039ca6334"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.720564 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4b54c868-49fn9" event={"ID":"c9233c70-4c77-41e8-831b-682307ae7dd7","Type":"ContainerStarted","Data":"9315e55b6a5b66d2bf3268e76993a3f5605fe6e6d5b38e3a8baca9469ed4beea"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.720617 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4b54c868-49fn9" event={"ID":"c9233c70-4c77-41e8-831b-682307ae7dd7","Type":"ContainerStarted","Data":"25c67a97f15a6015569ddccc90e85fc70cd306e5e61d6c52b3c6e8059e56a01a"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.720631 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d4b54c868-49fn9" event={"ID":"c9233c70-4c77-41e8-831b-682307ae7dd7","Type":"ContainerStarted","Data":"99181710d1ef5027908824ebce5fc158f1a123bb5522fc0467c5af655de83f8f"} Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.722325 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.723806 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-sb\") pod \"680cc550-1bfc-4a40-97e5-72067cb9efb7\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.723911 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-svc\") pod \"680cc550-1bfc-4a40-97e5-72067cb9efb7\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.723974 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-swift-storage-0\") pod \"680cc550-1bfc-4a40-97e5-72067cb9efb7\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.724052 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-nb\") pod \"680cc550-1bfc-4a40-97e5-72067cb9efb7\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.724182 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nl5q\" (UniqueName: \"kubernetes.io/projected/680cc550-1bfc-4a40-97e5-72067cb9efb7-kube-api-access-7nl5q\") pod \"680cc550-1bfc-4a40-97e5-72067cb9efb7\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.724260 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-config\") pod \"680cc550-1bfc-4a40-97e5-72067cb9efb7\" (UID: \"680cc550-1bfc-4a40-97e5-72067cb9efb7\") " Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.732739 4796 scope.go:117] "RemoveContainer" containerID="eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b" Nov 27 11:57:10 crc kubenswrapper[4796]: E1127 11:57:10.736746 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b\": container with ID starting with eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b not found: ID does not exist" containerID="eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.736796 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b"} err="failed to get container status \"eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b\": rpc error: code = NotFound desc = could not find container \"eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b\": container with ID starting with eb7f38c90d98d11d4f6e62509c1b23566c7e8c32a8de8365f472d7025675248b not found: ID does not exist" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.736840 4796 scope.go:117] "RemoveContainer" containerID="2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054" Nov 27 11:57:10 crc kubenswrapper[4796]: E1127 11:57:10.737314 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054\": container with ID starting with 2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054 not found: ID does not exist" containerID="2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.737340 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054"} err="failed to get container status \"2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054\": rpc error: code = NotFound desc = could not find container \"2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054\": container with ID starting with 2d1eeb04bf2461d0e1ca8d9b6b69e1b1e94eead4366b539cc5e9269467a53054 not found: ID does not exist" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.759543 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/680cc550-1bfc-4a40-97e5-72067cb9efb7-kube-api-access-7nl5q" (OuterVolumeSpecName: "kube-api-access-7nl5q") pod "680cc550-1bfc-4a40-97e5-72067cb9efb7" (UID: "680cc550-1bfc-4a40-97e5-72067cb9efb7"). InnerVolumeSpecName "kube-api-access-7nl5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.834973 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nl5q\" (UniqueName: \"kubernetes.io/projected/680cc550-1bfc-4a40-97e5-72067cb9efb7-kube-api-access-7nl5q\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.839373 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7d4b54c868-49fn9" podStartSLOduration=2.839353732 podStartE2EDuration="2.839353732s" podCreationTimestamp="2025-11-27 11:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:10.763676723 +0000 UTC m=+1948.281995651" watchObservedRunningTime="2025-11-27 11:57:10.839353732 +0000 UTC m=+1948.357672650" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.878414 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-config" (OuterVolumeSpecName: "config") pod "680cc550-1bfc-4a40-97e5-72067cb9efb7" (UID: "680cc550-1bfc-4a40-97e5-72067cb9efb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.900025 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "680cc550-1bfc-4a40-97e5-72067cb9efb7" (UID: "680cc550-1bfc-4a40-97e5-72067cb9efb7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.939542 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.939573 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:10 crc kubenswrapper[4796]: I1127 11:57:10.940444 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "680cc550-1bfc-4a40-97e5-72067cb9efb7" (UID: "680cc550-1bfc-4a40-97e5-72067cb9efb7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.039803 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "680cc550-1bfc-4a40-97e5-72067cb9efb7" (UID: "680cc550-1bfc-4a40-97e5-72067cb9efb7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.047761 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.047798 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.065655 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:11 crc kubenswrapper[4796]: E1127 11:57:11.066025 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerName="init" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066036 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerName="init" Nov 27 11:57:11 crc kubenswrapper[4796]: E1127 11:57:11.066051 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d550fe-17da-4b68-8e62-b537d6c07314" containerName="cinder-db-sync" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066059 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d550fe-17da-4b68-8e62-b537d6c07314" containerName="cinder-db-sync" Nov 27 11:57:11 crc kubenswrapper[4796]: E1127 11:57:11.066074 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerName="dnsmasq-dns" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066079 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerName="dnsmasq-dns" Nov 27 11:57:11 crc kubenswrapper[4796]: E1127 11:57:11.066100 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d920f619-249d-4266-ae4b-e13917ae78dc" containerName="heat-db-sync" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066106 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d920f619-249d-4266-ae4b-e13917ae78dc" containerName="heat-db-sync" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066289 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d550fe-17da-4b68-8e62-b537d6c07314" containerName="cinder-db-sync" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066300 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" containerName="dnsmasq-dns" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.066316 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d920f619-249d-4266-ae4b-e13917ae78dc" containerName="heat-db-sync" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.067242 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.080978 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "680cc550-1bfc-4a40-97e5-72067cb9efb7" (UID: "680cc550-1bfc-4a40-97e5-72067cb9efb7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.106673 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-vmg9q" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.122561 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.122730 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.122811 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150197 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59728\" (UniqueName: \"kubernetes.io/projected/8f4749d7-1edb-4830-aeeb-0ee60f975167-kube-api-access-59728\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150255 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150349 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150373 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150389 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f4749d7-1edb-4830-aeeb-0ee60f975167-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.150474 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/680cc550-1bfc-4a40-97e5-72067cb9efb7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.202978 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.288518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.288974 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.289101 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.289287 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.290806 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f4749d7-1edb-4830-aeeb-0ee60f975167-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.291199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59728\" (UniqueName: \"kubernetes.io/projected/8f4749d7-1edb-4830-aeeb-0ee60f975167-kube-api-access-59728\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.293955 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f4749d7-1edb-4830-aeeb-0ee60f975167-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.293866 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.319933 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.327480 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-27ssb"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.329873 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.342895 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.373538 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.378994 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59728\" (UniqueName: \"kubernetes.io/projected/8f4749d7-1edb-4830-aeeb-0ee60f975167-kube-api-access-59728\") pod \"cinder-scheduler-0\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.396312 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-svc\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.396408 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4wjk\" (UniqueName: \"kubernetes.io/projected/4f58631f-4476-46a4-969f-e6d7c989fb3e-kube-api-access-p4wjk\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.396441 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.396502 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.396627 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.396661 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-config\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.435748 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-27ssb"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.441334 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.463823 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.465457 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.468477 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.482281 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.497744 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-svc\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.497809 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4wjk\" (UniqueName: \"kubernetes.io/projected/4f58631f-4476-46a4-969f-e6d7c989fb3e-kube-api-access-p4wjk\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.497830 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.497875 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.497928 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.497946 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-config\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.498769 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-config\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.499404 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-svc\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.499980 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.500664 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.507518 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.521357 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4wjk\" (UniqueName: \"kubernetes.io/projected/4f58631f-4476-46a4-969f-e6d7c989fb3e-kube-api-access-p4wjk\") pod \"dnsmasq-dns-6578955fd5-27ssb\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599354 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599718 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbb25c70-ee95-4b50-bb15-b422c44b11a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599752 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-scripts\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599840 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5ggh\" (UniqueName: \"kubernetes.io/projected/dbb25c70-ee95-4b50-bb15-b422c44b11a2-kube-api-access-s5ggh\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599867 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb25c70-ee95-4b50-bb15-b422c44b11a2-logs\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599900 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.599917 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: E1127 11:57:11.616426 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod680cc550_1bfc_4a40_97e5_72067cb9efb7.slice/crio-697e421b63756cae1b525b76c4b1f66075f24b7d15a817f550581400aff96eec\": RecentStats: unable to find data in memory cache]" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.664430 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.701739 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-scripts\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.702579 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5ggh\" (UniqueName: \"kubernetes.io/projected/dbb25c70-ee95-4b50-bb15-b422c44b11a2-kube-api-access-s5ggh\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.702646 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb25c70-ee95-4b50-bb15-b422c44b11a2-logs\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.702714 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.702748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.702811 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.702905 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbb25c70-ee95-4b50-bb15-b422c44b11a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.703484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb25c70-ee95-4b50-bb15-b422c44b11a2-logs\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.703514 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbb25c70-ee95-4b50-bb15-b422c44b11a2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.706645 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.706913 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data-custom\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.707768 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-scripts\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.709777 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.717869 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5ggh\" (UniqueName: \"kubernetes.io/projected/dbb25c70-ee95-4b50-bb15-b422c44b11a2-kube-api-access-s5ggh\") pod \"cinder-api-0\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " pod="openstack/cinder-api-0" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.736232 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-mkwn4" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.736833 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.763605 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-mkwn4"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.775024 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-mkwn4"] Nov 27 11:57:11 crc kubenswrapper[4796]: I1127 11:57:11.790050 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 11:57:12 crc kubenswrapper[4796]: I1127 11:57:12.504004 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:12 crc kubenswrapper[4796]: I1127 11:57:12.569259 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:57:12 crc kubenswrapper[4796]: E1127 11:57:12.569514 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:57:12 crc kubenswrapper[4796]: I1127 11:57:12.637172 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:12 crc kubenswrapper[4796]: I1127 11:57:12.750465 4796 generic.go:334] "Generic (PLEG): container finished" podID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerID="295b69283f594d354574c7cb40ab077beea9846d5229d60a732e436dd705fa22" exitCode=0 Nov 27 11:57:12 crc kubenswrapper[4796]: I1127 11:57:12.750515 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerDied","Data":"295b69283f594d354574c7cb40ab077beea9846d5229d60a732e436dd705fa22"} Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.391891 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.530061 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-27ssb"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.548700 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-config-data\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.548845 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-combined-ca-bundle\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.548899 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-run-httpd\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.548925 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-sg-core-conf-yaml\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.548998 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-scripts\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.549031 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-log-httpd\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.549146 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b5jt\" (UniqueName: \"kubernetes.io/projected/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-kube-api-access-6b5jt\") pod \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\" (UID: \"13ee34c5-1762-4002-ba1e-63f3e1c90e6e\") " Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.550352 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.550518 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.550816 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.550841 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.556358 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-scripts" (OuterVolumeSpecName: "scripts") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.563412 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-kube-api-access-6b5jt" (OuterVolumeSpecName: "kube-api-access-6b5jt") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "kube-api-access-6b5jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.625234 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.629647 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="680cc550-1bfc-4a40-97e5-72067cb9efb7" path="/var/lib/kubelet/pods/680cc550-1bfc-4a40-97e5-72067cb9efb7/volumes" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.655365 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b5jt\" (UniqueName: \"kubernetes.io/projected/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-kube-api-access-6b5jt\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.655571 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.655580 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.680124 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.684030 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.697405 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.726700 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.731404 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-config-data" (OuterVolumeSpecName: "config-data") pod "13ee34c5-1762-4002-ba1e-63f3e1c90e6e" (UID: "13ee34c5-1762-4002-ba1e-63f3e1c90e6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.757963 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.757985 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13ee34c5-1762-4002-ba1e-63f3e1c90e6e-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.776727 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13ee34c5-1762-4002-ba1e-63f3e1c90e6e","Type":"ContainerDied","Data":"85614f297a6402221e4cd7ce5427483143e7b5667372fc4ddf917e3035313915"} Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.776796 4796 scope.go:117] "RemoveContainer" containerID="3c186646c11391054c8c86afc01f80accaeb1dccd4d12fdea3aa7e204727af4a" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.776743 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.784813 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f4749d7-1edb-4830-aeeb-0ee60f975167","Type":"ContainerStarted","Data":"e86b433d8568f1aa8047fb09da0c0bfcdeb3f0d107d140d023137edc9907554a"} Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.796751 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" event={"ID":"4f58631f-4476-46a4-969f-e6d7c989fb3e","Type":"ContainerStarted","Data":"6ea78b6de89439bf36325d2cd5ea143f6673e2c167d47d1070733062ac7c7048"} Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.801794 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbb25c70-ee95-4b50-bb15-b422c44b11a2","Type":"ContainerStarted","Data":"b8ca3ac2536bd40ac54c96437651969d360c9491e17e156979cafc2d90c84108"} Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.832631 4796 scope.go:117] "RemoveContainer" containerID="cb62072977045640d6230c54ad05fc56438fd69f2f79f3735f67ca9039ca6334" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.855333 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.864126 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.902871 4796 scope.go:117] "RemoveContainer" containerID="295b69283f594d354574c7cb40ab077beea9846d5229d60a732e436dd705fa22" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.902997 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: E1127 11:57:13.903401 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="sg-core" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.903418 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="sg-core" Nov 27 11:57:13 crc kubenswrapper[4796]: E1127 11:57:13.903431 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="ceilometer-notification-agent" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.903439 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="ceilometer-notification-agent" Nov 27 11:57:13 crc kubenswrapper[4796]: E1127 11:57:13.903466 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="proxy-httpd" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.903472 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="proxy-httpd" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.903644 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="sg-core" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.903661 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="proxy-httpd" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.903704 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" containerName="ceilometer-notification-agent" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.905371 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.908290 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.908534 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.912868 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966506 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-log-httpd\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966558 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966626 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966699 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-run-httpd\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966722 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt4qx\" (UniqueName: \"kubernetes.io/projected/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-kube-api-access-jt4qx\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966752 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-scripts\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:13 crc kubenswrapper[4796]: I1127 11:57:13.966786 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-config-data\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069225 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-run-httpd\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069300 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt4qx\" (UniqueName: \"kubernetes.io/projected/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-kube-api-access-jt4qx\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069341 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-scripts\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069379 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-config-data\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069472 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-log-httpd\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069494 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.069551 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.072254 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-run-httpd\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.073545 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-log-httpd\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.076047 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.077060 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-scripts\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.077647 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-config-data\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.089137 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.089431 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt4qx\" (UniqueName: \"kubernetes.io/projected/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-kube-api-access-jt4qx\") pod \"ceilometer-0\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.331818 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.810637 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.817722 4796 generic.go:334] "Generic (PLEG): container finished" podID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerID="b89f7f5e738b467e6b0a50a7f6a595d074233ecedb972e35279d6ebf2b5eb6ea" exitCode=0 Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.817793 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" event={"ID":"4f58631f-4476-46a4-969f-e6d7c989fb3e","Type":"ContainerDied","Data":"b89f7f5e738b467e6b0a50a7f6a595d074233ecedb972e35279d6ebf2b5eb6ea"} Nov 27 11:57:14 crc kubenswrapper[4796]: W1127 11:57:14.818726 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9482d56_62c4_41e8_b9ee_d5e9115cf8ac.slice/crio-7dc4eda8d03be1993e2e1653aff83f4a9264945ac27826c7259d7473abf2997e WatchSource:0}: Error finding container 7dc4eda8d03be1993e2e1653aff83f4a9264945ac27826c7259d7473abf2997e: Status 404 returned error can't find the container with id 7dc4eda8d03be1993e2e1653aff83f4a9264945ac27826c7259d7473abf2997e Nov 27 11:57:14 crc kubenswrapper[4796]: I1127 11:57:14.819906 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbb25c70-ee95-4b50-bb15-b422c44b11a2","Type":"ContainerStarted","Data":"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493"} Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.584350 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13ee34c5-1762-4002-ba1e-63f3e1c90e6e" path="/var/lib/kubelet/pods/13ee34c5-1762-4002-ba1e-63f3e1c90e6e/volumes" Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.833147 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerStarted","Data":"a1682de5b79877a5fa2df16f1765d759455dad04d13ee69478463532ac899deb"} Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.833202 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerStarted","Data":"7dc4eda8d03be1993e2e1653aff83f4a9264945ac27826c7259d7473abf2997e"} Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.835115 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f4749d7-1edb-4830-aeeb-0ee60f975167","Type":"ContainerStarted","Data":"9aa44b92325fea59f9471b563457d8bd21dbde8f044b39f8503db6692a191990"} Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.838110 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" event={"ID":"4f58631f-4476-46a4-969f-e6d7c989fb3e","Type":"ContainerStarted","Data":"b4b7555a272a493026f39d1be9a5119c131b08da4f9989e0317c66e9eb6a6850"} Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.839341 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.842593 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbb25c70-ee95-4b50-bb15-b422c44b11a2","Type":"ContainerStarted","Data":"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025"} Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.842712 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api-log" containerID="cri-o://a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493" gracePeriod=30 Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.842814 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.842823 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api" containerID="cri-o://627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025" gracePeriod=30 Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.875158 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" podStartSLOduration=4.875132841 podStartE2EDuration="4.875132841s" podCreationTimestamp="2025-11-27 11:57:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:15.861822396 +0000 UTC m=+1953.380141324" watchObservedRunningTime="2025-11-27 11:57:15.875132841 +0000 UTC m=+1953.393451759" Nov 27 11:57:15 crc kubenswrapper[4796]: I1127 11:57:15.907334 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.907312589 podStartE2EDuration="4.907312589s" podCreationTimestamp="2025-11-27 11:57:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:15.889042442 +0000 UTC m=+1953.407361360" watchObservedRunningTime="2025-11-27 11:57:15.907312589 +0000 UTC m=+1953.425631507" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.503166 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634310 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb25c70-ee95-4b50-bb15-b422c44b11a2-logs\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634376 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbb25c70-ee95-4b50-bb15-b422c44b11a2-etc-machine-id\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634461 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5ggh\" (UniqueName: \"kubernetes.io/projected/dbb25c70-ee95-4b50-bb15-b422c44b11a2-kube-api-access-s5ggh\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634476 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dbb25c70-ee95-4b50-bb15-b422c44b11a2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634541 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data-custom\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634611 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634749 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-scripts\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634767 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbb25c70-ee95-4b50-bb15-b422c44b11a2-logs" (OuterVolumeSpecName: "logs") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.634813 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-combined-ca-bundle\") pod \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\" (UID: \"dbb25c70-ee95-4b50-bb15-b422c44b11a2\") " Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.635394 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb25c70-ee95-4b50-bb15-b422c44b11a2-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.635420 4796 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbb25c70-ee95-4b50-bb15-b422c44b11a2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.640404 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.640619 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbb25c70-ee95-4b50-bb15-b422c44b11a2-kube-api-access-s5ggh" (OuterVolumeSpecName: "kube-api-access-s5ggh") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "kube-api-access-s5ggh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.640792 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-scripts" (OuterVolumeSpecName: "scripts") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.671406 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.707654 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data" (OuterVolumeSpecName: "config-data") pod "dbb25c70-ee95-4b50-bb15-b422c44b11a2" (UID: "dbb25c70-ee95-4b50-bb15-b422c44b11a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.736999 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5ggh\" (UniqueName: \"kubernetes.io/projected/dbb25c70-ee95-4b50-bb15-b422c44b11a2-kube-api-access-s5ggh\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.737749 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.737762 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.737773 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.737785 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb25c70-ee95-4b50-bb15-b422c44b11a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854371 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerID="627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025" exitCode=0 Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854406 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerID="a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493" exitCode=143 Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854453 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbb25c70-ee95-4b50-bb15-b422c44b11a2","Type":"ContainerDied","Data":"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025"} Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854483 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbb25c70-ee95-4b50-bb15-b422c44b11a2","Type":"ContainerDied","Data":"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493"} Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854498 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbb25c70-ee95-4b50-bb15-b422c44b11a2","Type":"ContainerDied","Data":"b8ca3ac2536bd40ac54c96437651969d360c9491e17e156979cafc2d90c84108"} Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854518 4796 scope.go:117] "RemoveContainer" containerID="627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.854647 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.860142 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerStarted","Data":"531045a7a9b35b5292ea4d0d2a253bc76988ac519e3a38e9362e097da7a87f2e"} Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.863343 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f4749d7-1edb-4830-aeeb-0ee60f975167","Type":"ContainerStarted","Data":"59951ef5d92a781dfd107c94b848f2d0dca5886a40527203d68852b5f0a386fa"} Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.907181 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.789512487 podStartE2EDuration="6.907157252s" podCreationTimestamp="2025-11-27 11:57:10 +0000 UTC" firstStartedPulling="2025-11-27 11:57:13.702322067 +0000 UTC m=+1951.220640985" lastFinishedPulling="2025-11-27 11:57:14.819966832 +0000 UTC m=+1952.338285750" observedRunningTime="2025-11-27 11:57:16.881257642 +0000 UTC m=+1954.399576580" watchObservedRunningTime="2025-11-27 11:57:16.907157252 +0000 UTC m=+1954.425476170" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.934540 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.939391 4796 scope.go:117] "RemoveContainer" containerID="a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.954623 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.962976 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:16 crc kubenswrapper[4796]: E1127 11:57:16.963826 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api-log" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.963861 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api-log" Nov 27 11:57:16 crc kubenswrapper[4796]: E1127 11:57:16.963879 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.963885 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.964133 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.964170 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" containerName="cinder-api-log" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.966925 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.974695 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.976337 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.976536 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 27 11:57:16 crc kubenswrapper[4796]: I1127 11:57:16.976643 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.017879 4796 scope.go:117] "RemoveContainer" containerID="627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025" Nov 27 11:57:17 crc kubenswrapper[4796]: E1127 11:57:17.018302 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025\": container with ID starting with 627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025 not found: ID does not exist" containerID="627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.018331 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025"} err="failed to get container status \"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025\": rpc error: code = NotFound desc = could not find container \"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025\": container with ID starting with 627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025 not found: ID does not exist" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.018371 4796 scope.go:117] "RemoveContainer" containerID="a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493" Nov 27 11:57:17 crc kubenswrapper[4796]: E1127 11:57:17.018584 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493\": container with ID starting with a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493 not found: ID does not exist" containerID="a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.018632 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493"} err="failed to get container status \"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493\": rpc error: code = NotFound desc = could not find container \"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493\": container with ID starting with a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493 not found: ID does not exist" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.018649 4796 scope.go:117] "RemoveContainer" containerID="627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.018887 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025"} err="failed to get container status \"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025\": rpc error: code = NotFound desc = could not find container \"627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025\": container with ID starting with 627bffdb2af0f845a97989b616ba00273ac56b7d280f9e7541f24186263d5025 not found: ID does not exist" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.018943 4796 scope.go:117] "RemoveContainer" containerID="a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.019217 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493"} err="failed to get container status \"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493\": rpc error: code = NotFound desc = could not find container \"a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493\": container with ID starting with a33619ebef13558e7d25b3b56efa2aa20291e7b959b29504672b91ad2af9c493 not found: ID does not exist" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042445 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae7bbf2-68f9-489d-ab84-37202429a5fb-logs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042485 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042506 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042537 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ae7bbf2-68f9-489d-ab84-37202429a5fb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-scripts\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042594 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shslv\" (UniqueName: \"kubernetes.io/projected/7ae7bbf2-68f9-489d-ab84-37202429a5fb-kube-api-access-shslv\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042637 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-config-data\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.042668 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-config-data-custom\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144311 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-scripts\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144373 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shslv\" (UniqueName: \"kubernetes.io/projected/7ae7bbf2-68f9-489d-ab84-37202429a5fb-kube-api-access-shslv\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144400 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144420 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-config-data\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144455 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-config-data-custom\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae7bbf2-68f9-489d-ab84-37202429a5fb-logs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144538 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144554 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144582 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ae7bbf2-68f9-489d-ab84-37202429a5fb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.144659 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7ae7bbf2-68f9-489d-ab84-37202429a5fb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.146200 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae7bbf2-68f9-489d-ab84-37202429a5fb-logs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.165137 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.165907 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.165992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-scripts\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.168289 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-config-data-custom\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.178656 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-config-data\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.179294 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae7bbf2-68f9-489d-ab84-37202429a5fb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.185980 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shslv\" (UniqueName: \"kubernetes.io/projected/7ae7bbf2-68f9-489d-ab84-37202429a5fb-kube-api-access-shslv\") pod \"cinder-api-0\" (UID: \"7ae7bbf2-68f9-489d-ab84-37202429a5fb\") " pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.311498 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.581290 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbb25c70-ee95-4b50-bb15-b422c44b11a2" path="/var/lib/kubelet/pods/dbb25c70-ee95-4b50-bb15-b422c44b11a2/volumes" Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.747073 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 11:57:17 crc kubenswrapper[4796]: W1127 11:57:17.749020 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ae7bbf2_68f9_489d_ab84_37202429a5fb.slice/crio-80e8baffe7121b990e67562794e4416cac73419aab27a12d19e635c8b5684e13 WatchSource:0}: Error finding container 80e8baffe7121b990e67562794e4416cac73419aab27a12d19e635c8b5684e13: Status 404 returned error can't find the container with id 80e8baffe7121b990e67562794e4416cac73419aab27a12d19e635c8b5684e13 Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.874260 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7ae7bbf2-68f9-489d-ab84-37202429a5fb","Type":"ContainerStarted","Data":"80e8baffe7121b990e67562794e4416cac73419aab27a12d19e635c8b5684e13"} Nov 27 11:57:17 crc kubenswrapper[4796]: I1127 11:57:17.880395 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerStarted","Data":"397181c8cad6014c0dcf698f94e281712dc9ca36a9fedc8036933831e64e2977"} Nov 27 11:57:18 crc kubenswrapper[4796]: I1127 11:57:18.895034 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7ae7bbf2-68f9-489d-ab84-37202429a5fb","Type":"ContainerStarted","Data":"e6063603c4e70dd815b2dadf3c2fe78aeb3c9f7e5d869137f3f1429326711a1a"} Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.663285 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.769717 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.910309 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerStarted","Data":"2be13e35bc18b161ce897b5509527b181d5f746119d5d48ed7ae2c70e25433a3"} Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.910393 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.911992 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7ae7bbf2-68f9-489d-ab84-37202429a5fb","Type":"ContainerStarted","Data":"540a4cdda5ce002aa36eb0ec8b5281a4df3690bd114434efa57eb90b5486e27c"} Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.912110 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.943550 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.540260687 podStartE2EDuration="6.943529513s" podCreationTimestamp="2025-11-27 11:57:13 +0000 UTC" firstStartedPulling="2025-11-27 11:57:14.823198788 +0000 UTC m=+1952.341517716" lastFinishedPulling="2025-11-27 11:57:19.226467614 +0000 UTC m=+1956.744786542" observedRunningTime="2025-11-27 11:57:19.940645686 +0000 UTC m=+1957.458964614" watchObservedRunningTime="2025-11-27 11:57:19.943529513 +0000 UTC m=+1957.461848431" Nov 27 11:57:19 crc kubenswrapper[4796]: I1127 11:57:19.961948 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.961924494 podStartE2EDuration="3.961924494s" podCreationTimestamp="2025-11-27 11:57:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:19.957703391 +0000 UTC m=+1957.476022329" watchObservedRunningTime="2025-11-27 11:57:19.961924494 +0000 UTC m=+1957.480243422" Nov 27 11:57:20 crc kubenswrapper[4796]: I1127 11:57:20.072929 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-655d9778b6-79p2x" Nov 27 11:57:20 crc kubenswrapper[4796]: I1127 11:57:20.126155 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5dcd9f4978-tl45d"] Nov 27 11:57:20 crc kubenswrapper[4796]: I1127 11:57:20.127899 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5dcd9f4978-tl45d" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api" containerID="cri-o://5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c" gracePeriod=30 Nov 27 11:57:20 crc kubenswrapper[4796]: I1127 11:57:20.131214 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5dcd9f4978-tl45d" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api-log" containerID="cri-o://8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a" gracePeriod=30 Nov 27 11:57:20 crc kubenswrapper[4796]: I1127 11:57:20.922729 4796 generic.go:334] "Generic (PLEG): container finished" podID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerID="8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a" exitCode=143 Nov 27 11:57:20 crc kubenswrapper[4796]: I1127 11:57:20.922795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dcd9f4978-tl45d" event={"ID":"91118dbe-a3ca-41b6-ab86-a24ebd7b5321","Type":"ContainerDied","Data":"8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a"} Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.442862 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.663694 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.667050 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.785050 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zwrc6"] Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.785571 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerName="dnsmasq-dns" containerID="cri-o://20bfd9fd2b3fc49622ee7161c967ac41cf3b51caedaf10afa53ab77b17dd9043" gracePeriod=10 Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.946576 4796 generic.go:334] "Generic (PLEG): container finished" podID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerID="20bfd9fd2b3fc49622ee7161c967ac41cf3b51caedaf10afa53ab77b17dd9043" exitCode=0 Nov 27 11:57:21 crc kubenswrapper[4796]: I1127 11:57:21.946660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" event={"ID":"1a852724-a4bb-4ed6-aa8e-da99f2d31de2","Type":"ContainerDied","Data":"20bfd9fd2b3fc49622ee7161c967ac41cf3b51caedaf10afa53ab77b17dd9043"} Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.003246 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.039697 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69bc59f7f-g45f2" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.108945 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55ffbfd4fd-lldkk"] Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.109226 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55ffbfd4fd-lldkk" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-api" containerID="cri-o://dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc" gracePeriod=30 Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.109397 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55ffbfd4fd-lldkk" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-httpd" containerID="cri-o://77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7" gracePeriod=30 Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.335437 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.473426 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-config\") pod \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.473518 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-nb\") pod \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.473554 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-swift-storage-0\") pod \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.473620 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fccgc\" (UniqueName: \"kubernetes.io/projected/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-kube-api-access-fccgc\") pod \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.473692 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-svc\") pod \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.473781 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-sb\") pod \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\" (UID: \"1a852724-a4bb-4ed6-aa8e-da99f2d31de2\") " Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.483472 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-kube-api-access-fccgc" (OuterVolumeSpecName: "kube-api-access-fccgc") pod "1a852724-a4bb-4ed6-aa8e-da99f2d31de2" (UID: "1a852724-a4bb-4ed6-aa8e-da99f2d31de2"). InnerVolumeSpecName "kube-api-access-fccgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.536948 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1a852724-a4bb-4ed6-aa8e-da99f2d31de2" (UID: "1a852724-a4bb-4ed6-aa8e-da99f2d31de2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.538943 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1a852724-a4bb-4ed6-aa8e-da99f2d31de2" (UID: "1a852724-a4bb-4ed6-aa8e-da99f2d31de2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.551691 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-config" (OuterVolumeSpecName: "config") pod "1a852724-a4bb-4ed6-aa8e-da99f2d31de2" (UID: "1a852724-a4bb-4ed6-aa8e-da99f2d31de2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.563502 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1a852724-a4bb-4ed6-aa8e-da99f2d31de2" (UID: "1a852724-a4bb-4ed6-aa8e-da99f2d31de2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.567887 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1a852724-a4bb-4ed6-aa8e-da99f2d31de2" (UID: "1a852724-a4bb-4ed6-aa8e-da99f2d31de2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.575645 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.575692 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.575701 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.575709 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.575721 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fccgc\" (UniqueName: \"kubernetes.io/projected/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-kube-api-access-fccgc\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.575731 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a852724-a4bb-4ed6-aa8e-da99f2d31de2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.960198 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" event={"ID":"1a852724-a4bb-4ed6-aa8e-da99f2d31de2","Type":"ContainerDied","Data":"7659b6edf86ced3a07c3b826a82c5a234fa81610f97652da66d43e603802a7fc"} Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.960239 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-zwrc6" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.960302 4796 scope.go:117] "RemoveContainer" containerID="20bfd9fd2b3fc49622ee7161c967ac41cf3b51caedaf10afa53ab77b17dd9043" Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.963573 4796 generic.go:334] "Generic (PLEG): container finished" podID="4eb23259-99e7-4823-8820-54eb455ae19f" containerID="77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7" exitCode=0 Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.963735 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55ffbfd4fd-lldkk" event={"ID":"4eb23259-99e7-4823-8820-54eb455ae19f","Type":"ContainerDied","Data":"77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7"} Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.963842 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="cinder-scheduler" containerID="cri-o://9aa44b92325fea59f9471b563457d8bd21dbde8f044b39f8503db6692a191990" gracePeriod=30 Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.964018 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="probe" containerID="cri-o://59951ef5d92a781dfd107c94b848f2d0dca5886a40527203d68852b5f0a386fa" gracePeriod=30 Nov 27 11:57:22 crc kubenswrapper[4796]: I1127 11:57:22.988401 4796 scope.go:117] "RemoveContainer" containerID="e6bf0c7fc2f34a7e1c3d14584a6a358efcd59b6f258bbddeb005e0fe737f0db2" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.007413 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zwrc6"] Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.016412 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-zwrc6"] Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.583073 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" path="/var/lib/kubelet/pods/1a852724-a4bb-4ed6-aa8e-da99f2d31de2/volumes" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.700707 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.796564 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data-custom\") pod \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.796732 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-logs\") pod \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.796834 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-combined-ca-bundle\") pod \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.796881 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v27vq\" (UniqueName: \"kubernetes.io/projected/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-kube-api-access-v27vq\") pod \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.797016 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data\") pod \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\" (UID: \"91118dbe-a3ca-41b6-ab86-a24ebd7b5321\") " Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.797797 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-logs" (OuterVolumeSpecName: "logs") pod "91118dbe-a3ca-41b6-ab86-a24ebd7b5321" (UID: "91118dbe-a3ca-41b6-ab86-a24ebd7b5321"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.802057 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "91118dbe-a3ca-41b6-ab86-a24ebd7b5321" (UID: "91118dbe-a3ca-41b6-ab86-a24ebd7b5321"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.802569 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-kube-api-access-v27vq" (OuterVolumeSpecName: "kube-api-access-v27vq") pod "91118dbe-a3ca-41b6-ab86-a24ebd7b5321" (UID: "91118dbe-a3ca-41b6-ab86-a24ebd7b5321"). InnerVolumeSpecName "kube-api-access-v27vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.828935 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91118dbe-a3ca-41b6-ab86-a24ebd7b5321" (UID: "91118dbe-a3ca-41b6-ab86-a24ebd7b5321"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.854826 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data" (OuterVolumeSpecName: "config-data") pod "91118dbe-a3ca-41b6-ab86-a24ebd7b5321" (UID: "91118dbe-a3ca-41b6-ab86-a24ebd7b5321"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.899370 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.899409 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v27vq\" (UniqueName: \"kubernetes.io/projected/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-kube-api-access-v27vq\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.899426 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.899439 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.899453 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91118dbe-a3ca-41b6-ab86-a24ebd7b5321-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.986731 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerID="59951ef5d92a781dfd107c94b848f2d0dca5886a40527203d68852b5f0a386fa" exitCode=0 Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.986818 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f4749d7-1edb-4830-aeeb-0ee60f975167","Type":"ContainerDied","Data":"59951ef5d92a781dfd107c94b848f2d0dca5886a40527203d68852b5f0a386fa"} Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.990584 4796 generic.go:334] "Generic (PLEG): container finished" podID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerID="5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c" exitCode=0 Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.990637 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dcd9f4978-tl45d" event={"ID":"91118dbe-a3ca-41b6-ab86-a24ebd7b5321","Type":"ContainerDied","Data":"5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c"} Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.990658 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dcd9f4978-tl45d" event={"ID":"91118dbe-a3ca-41b6-ab86-a24ebd7b5321","Type":"ContainerDied","Data":"129d7511e21e2e3d15660f4d0d24573627d437820fe521be88fcfdda9392addd"} Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.990678 4796 scope.go:117] "RemoveContainer" containerID="5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c" Nov 27 11:57:23 crc kubenswrapper[4796]: I1127 11:57:23.990787 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dcd9f4978-tl45d" Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.035932 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5dcd9f4978-tl45d"] Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.045053 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5dcd9f4978-tl45d"] Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.060746 4796 scope.go:117] "RemoveContainer" containerID="8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a" Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.083579 4796 scope.go:117] "RemoveContainer" containerID="5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c" Nov 27 11:57:24 crc kubenswrapper[4796]: E1127 11:57:24.084151 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c\": container with ID starting with 5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c not found: ID does not exist" containerID="5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c" Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.084190 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c"} err="failed to get container status \"5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c\": rpc error: code = NotFound desc = could not find container \"5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c\": container with ID starting with 5490316d65784500dec681db9d6b87402fab3417529e287067658a734daf958c not found: ID does not exist" Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.084217 4796 scope.go:117] "RemoveContainer" containerID="8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a" Nov 27 11:57:24 crc kubenswrapper[4796]: E1127 11:57:24.084649 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a\": container with ID starting with 8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a not found: ID does not exist" containerID="8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a" Nov 27 11:57:24 crc kubenswrapper[4796]: I1127 11:57:24.084679 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a"} err="failed to get container status \"8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a\": rpc error: code = NotFound desc = could not find container \"8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a\": container with ID starting with 8c34e63d57e1444f4fcb4b0250a351825e8b76fbf5f3cc124bf81a526001ee5a not found: ID does not exist" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.578427 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" path="/var/lib/kubelet/pods/91118dbe-a3ca-41b6-ab86-a24ebd7b5321/volumes" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.592567 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.633725 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-httpd-config\") pod \"4eb23259-99e7-4823-8820-54eb455ae19f\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.633822 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-config\") pod \"4eb23259-99e7-4823-8820-54eb455ae19f\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.633913 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-ovndb-tls-certs\") pod \"4eb23259-99e7-4823-8820-54eb455ae19f\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.633983 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-combined-ca-bundle\") pod \"4eb23259-99e7-4823-8820-54eb455ae19f\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.634111 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsvfd\" (UniqueName: \"kubernetes.io/projected/4eb23259-99e7-4823-8820-54eb455ae19f-kube-api-access-zsvfd\") pod \"4eb23259-99e7-4823-8820-54eb455ae19f\" (UID: \"4eb23259-99e7-4823-8820-54eb455ae19f\") " Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.642394 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4eb23259-99e7-4823-8820-54eb455ae19f" (UID: "4eb23259-99e7-4823-8820-54eb455ae19f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.648107 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb23259-99e7-4823-8820-54eb455ae19f-kube-api-access-zsvfd" (OuterVolumeSpecName: "kube-api-access-zsvfd") pod "4eb23259-99e7-4823-8820-54eb455ae19f" (UID: "4eb23259-99e7-4823-8820-54eb455ae19f"). InnerVolumeSpecName "kube-api-access-zsvfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.689779 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4eb23259-99e7-4823-8820-54eb455ae19f" (UID: "4eb23259-99e7-4823-8820-54eb455ae19f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.694500 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-config" (OuterVolumeSpecName: "config") pod "4eb23259-99e7-4823-8820-54eb455ae19f" (UID: "4eb23259-99e7-4823-8820-54eb455ae19f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.714839 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4eb23259-99e7-4823-8820-54eb455ae19f" (UID: "4eb23259-99e7-4823-8820-54eb455ae19f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.736427 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.736472 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.736485 4796 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.736501 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eb23259-99e7-4823-8820-54eb455ae19f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:25 crc kubenswrapper[4796]: I1127 11:57:25.736512 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsvfd\" (UniqueName: \"kubernetes.io/projected/4eb23259-99e7-4823-8820-54eb455ae19f-kube-api-access-zsvfd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.029305 4796 generic.go:334] "Generic (PLEG): container finished" podID="4eb23259-99e7-4823-8820-54eb455ae19f" containerID="dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc" exitCode=0 Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.029356 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55ffbfd4fd-lldkk" event={"ID":"4eb23259-99e7-4823-8820-54eb455ae19f","Type":"ContainerDied","Data":"dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc"} Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.029394 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55ffbfd4fd-lldkk" event={"ID":"4eb23259-99e7-4823-8820-54eb455ae19f","Type":"ContainerDied","Data":"75f5e8df808ed56133663fb2788267e1c020b80879e5d5c3222ca9f6428cacae"} Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.029393 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55ffbfd4fd-lldkk" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.029415 4796 scope.go:117] "RemoveContainer" containerID="77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.055302 4796 scope.go:117] "RemoveContainer" containerID="dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.069861 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55ffbfd4fd-lldkk"] Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.078954 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-55ffbfd4fd-lldkk"] Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.081895 4796 scope.go:117] "RemoveContainer" containerID="77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7" Nov 27 11:57:26 crc kubenswrapper[4796]: E1127 11:57:26.082400 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7\": container with ID starting with 77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7 not found: ID does not exist" containerID="77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.082433 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7"} err="failed to get container status \"77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7\": rpc error: code = NotFound desc = could not find container \"77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7\": container with ID starting with 77e7f4643cffd01b8be469549b843c663f5cf90c66df2764c384c99b1bb534c7 not found: ID does not exist" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.082464 4796 scope.go:117] "RemoveContainer" containerID="dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc" Nov 27 11:57:26 crc kubenswrapper[4796]: E1127 11:57:26.082779 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc\": container with ID starting with dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc not found: ID does not exist" containerID="dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.082800 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc"} err="failed to get container status \"dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc\": rpc error: code = NotFound desc = could not find container \"dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc\": container with ID starting with dde0f435811e7a9e48e2ee4849dfbdc7d9bafac592526d090d8dd3e709a986bc not found: ID does not exist" Nov 27 11:57:26 crc kubenswrapper[4796]: I1127 11:57:26.570020 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:57:26 crc kubenswrapper[4796]: E1127 11:57:26.570879 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:57:27 crc kubenswrapper[4796]: I1127 11:57:27.579559 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" path="/var/lib/kubelet/pods/4eb23259-99e7-4823-8820-54eb455ae19f/volumes" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.070904 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerID="9aa44b92325fea59f9471b563457d8bd21dbde8f044b39f8503db6692a191990" exitCode=0 Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.070954 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f4749d7-1edb-4830-aeeb-0ee60f975167","Type":"ContainerDied","Data":"9aa44b92325fea59f9471b563457d8bd21dbde8f044b39f8503db6692a191990"} Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.639124 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.688828 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data-custom\") pod \"8f4749d7-1edb-4830-aeeb-0ee60f975167\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.688919 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f4749d7-1edb-4830-aeeb-0ee60f975167-etc-machine-id\") pod \"8f4749d7-1edb-4830-aeeb-0ee60f975167\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.689029 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8f4749d7-1edb-4830-aeeb-0ee60f975167-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8f4749d7-1edb-4830-aeeb-0ee60f975167" (UID: "8f4749d7-1edb-4830-aeeb-0ee60f975167"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.689075 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data\") pod \"8f4749d7-1edb-4830-aeeb-0ee60f975167\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.689180 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59728\" (UniqueName: \"kubernetes.io/projected/8f4749d7-1edb-4830-aeeb-0ee60f975167-kube-api-access-59728\") pod \"8f4749d7-1edb-4830-aeeb-0ee60f975167\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.689208 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-scripts\") pod \"8f4749d7-1edb-4830-aeeb-0ee60f975167\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.689244 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-combined-ca-bundle\") pod \"8f4749d7-1edb-4830-aeeb-0ee60f975167\" (UID: \"8f4749d7-1edb-4830-aeeb-0ee60f975167\") " Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.690644 4796 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f4749d7-1edb-4830-aeeb-0ee60f975167-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.695717 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-scripts" (OuterVolumeSpecName: "scripts") pod "8f4749d7-1edb-4830-aeeb-0ee60f975167" (UID: "8f4749d7-1edb-4830-aeeb-0ee60f975167"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.695798 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8f4749d7-1edb-4830-aeeb-0ee60f975167" (UID: "8f4749d7-1edb-4830-aeeb-0ee60f975167"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.696162 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f4749d7-1edb-4830-aeeb-0ee60f975167-kube-api-access-59728" (OuterVolumeSpecName: "kube-api-access-59728") pod "8f4749d7-1edb-4830-aeeb-0ee60f975167" (UID: "8f4749d7-1edb-4830-aeeb-0ee60f975167"). InnerVolumeSpecName "kube-api-access-59728". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.756436 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f4749d7-1edb-4830-aeeb-0ee60f975167" (UID: "8f4749d7-1edb-4830-aeeb-0ee60f975167"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.785329 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data" (OuterVolumeSpecName: "config-data") pod "8f4749d7-1edb-4830-aeeb-0ee60f975167" (UID: "8f4749d7-1edb-4830-aeeb-0ee60f975167"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.792507 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.792534 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.792542 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.792551 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59728\" (UniqueName: \"kubernetes.io/projected/8f4749d7-1edb-4830-aeeb-0ee60f975167-kube-api-access-59728\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:28 crc kubenswrapper[4796]: I1127 11:57:28.792561 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f4749d7-1edb-4830-aeeb-0ee60f975167-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.080828 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f4749d7-1edb-4830-aeeb-0ee60f975167","Type":"ContainerDied","Data":"e86b433d8568f1aa8047fb09da0c0bfcdeb3f0d107d140d023137edc9907554a"} Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.081146 4796 scope.go:117] "RemoveContainer" containerID="59951ef5d92a781dfd107c94b848f2d0dca5886a40527203d68852b5f0a386fa" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.081294 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.119186 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.121501 4796 scope.go:117] "RemoveContainer" containerID="9aa44b92325fea59f9471b563457d8bd21dbde8f044b39f8503db6692a191990" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.122842 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.128538 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.160863 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161319 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerName="init" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161337 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerName="init" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161351 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-api" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161358 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-api" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161369 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api-log" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161375 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api-log" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161400 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="probe" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161405 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="probe" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161413 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="cinder-scheduler" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161420 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="cinder-scheduler" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161432 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161438 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161448 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerName="dnsmasq-dns" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161454 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerName="dnsmasq-dns" Nov 27 11:57:29 crc kubenswrapper[4796]: E1127 11:57:29.161469 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-httpd" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161475 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-httpd" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161649 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="probe" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161663 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-httpd" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161673 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb23259-99e7-4823-8820-54eb455ae19f" containerName="neutron-api" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161680 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" containerName="cinder-scheduler" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161693 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a852724-a4bb-4ed6-aa8e-da99f2d31de2" containerName="dnsmasq-dns" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161705 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.161713 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="91118dbe-a3ca-41b6-ab86-a24ebd7b5321" containerName="barbican-api-log" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.162661 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.169172 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.184472 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.198651 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vphfl\" (UniqueName: \"kubernetes.io/projected/f8373371-1f42-4460-8724-668392be8975-kube-api-access-vphfl\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.198751 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.198773 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.198863 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8373371-1f42-4460-8724-668392be8975-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.198880 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.198911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300446 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300489 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8373371-1f42-4460-8724-668392be8975-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300579 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300600 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300648 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vphfl\" (UniqueName: \"kubernetes.io/projected/f8373371-1f42-4460-8724-668392be8975-kube-api-access-vphfl\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.300701 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8373371-1f42-4460-8724-668392be8975-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.306175 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.306305 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.306574 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.306679 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8373371-1f42-4460-8724-668392be8975-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.319119 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vphfl\" (UniqueName: \"kubernetes.io/projected/f8373371-1f42-4460-8724-668392be8975-kube-api-access-vphfl\") pod \"cinder-scheduler-0\" (UID: \"f8373371-1f42-4460-8724-668392be8975\") " pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.480848 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.591513 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f4749d7-1edb-4830-aeeb-0ee60f975167" path="/var/lib/kubelet/pods/8f4749d7-1edb-4830-aeeb-0ee60f975167/volumes" Nov 27 11:57:29 crc kubenswrapper[4796]: I1127 11:57:29.968887 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 11:57:30 crc kubenswrapper[4796]: I1127 11:57:30.097087 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8373371-1f42-4460-8724-668392be8975","Type":"ContainerStarted","Data":"7ec1fcec10548d00356b2b50b0ebf504ae3750557cca7a2cd0c43fa46e0d889b"} Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.133782 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8373371-1f42-4460-8724-668392be8975","Type":"ContainerStarted","Data":"d3c20e71be37b5d2e453f2e4758aaada766e92c7613ebee81b6ce35c56c503c2"} Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.200742 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-95c9f9f9b-zshfm" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.925964 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.932299 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.934287 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.936233 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.936441 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xh8jt" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.936541 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.967168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.967301 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zclr5\" (UniqueName: \"kubernetes.io/projected/c54eb17e-bce8-4e91-b9f2-766040bcfc89-kube-api-access-zclr5\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.967348 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config-secret\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:31 crc kubenswrapper[4796]: I1127 11:57:31.967425 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.068905 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config-secret\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.069022 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.069183 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.069288 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zclr5\" (UniqueName: \"kubernetes.io/projected/c54eb17e-bce8-4e91-b9f2-766040bcfc89-kube-api-access-zclr5\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.070504 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.076678 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.078727 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config-secret\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.086507 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zclr5\" (UniqueName: \"kubernetes.io/projected/c54eb17e-bce8-4e91-b9f2-766040bcfc89-kube-api-access-zclr5\") pod \"openstackclient\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.145136 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8373371-1f42-4460-8724-668392be8975","Type":"ContainerStarted","Data":"b072ec96a7cbb6335f6cf6c684729c8939fc2555eb8b3d13f0dc466c0b920302"} Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.168854 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.168836816 podStartE2EDuration="3.168836816s" podCreationTimestamp="2025-11-27 11:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:32.161024697 +0000 UTC m=+1969.679343615" watchObservedRunningTime="2025-11-27 11:57:32.168836816 +0000 UTC m=+1969.687155734" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.262632 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 11:57:32 crc kubenswrapper[4796]: I1127 11:57:32.724480 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 11:57:32 crc kubenswrapper[4796]: W1127 11:57:32.737868 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc54eb17e_bce8_4e91_b9f2_766040bcfc89.slice/crio-1f32ef985249bbdfc08434c2541f3e6bbab5ce6bf1dbc868111e151ec082ee4e WatchSource:0}: Error finding container 1f32ef985249bbdfc08434c2541f3e6bbab5ce6bf1dbc868111e151ec082ee4e: Status 404 returned error can't find the container with id 1f32ef985249bbdfc08434c2541f3e6bbab5ce6bf1dbc868111e151ec082ee4e Nov 27 11:57:33 crc kubenswrapper[4796]: I1127 11:57:33.154245 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c54eb17e-bce8-4e91-b9f2-766040bcfc89","Type":"ContainerStarted","Data":"1f32ef985249bbdfc08434c2541f3e6bbab5ce6bf1dbc868111e151ec082ee4e"} Nov 27 11:57:34 crc kubenswrapper[4796]: I1127 11:57:34.481125 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.283498 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6568564dcc-mst5g"] Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.285172 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.288564 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.288872 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.296229 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.308177 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6568564dcc-mst5g"] Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325124 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bjb7\" (UniqueName: \"kubernetes.io/projected/c65b890a-db52-4342-938c-d56538e76b3c-kube-api-access-9bjb7\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325227 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65b890a-db52-4342-938c-d56538e76b3c-run-httpd\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325308 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-config-data\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325402 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c65b890a-db52-4342-938c-d56538e76b3c-etc-swift\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325434 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-internal-tls-certs\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325467 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-combined-ca-bundle\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325578 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-public-tls-certs\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.325626 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65b890a-db52-4342-938c-d56538e76b3c-log-httpd\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427028 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c65b890a-db52-4342-938c-d56538e76b3c-etc-swift\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427075 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-internal-tls-certs\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427093 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-combined-ca-bundle\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427155 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-public-tls-certs\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427185 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65b890a-db52-4342-938c-d56538e76b3c-log-httpd\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427201 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bjb7\" (UniqueName: \"kubernetes.io/projected/c65b890a-db52-4342-938c-d56538e76b3c-kube-api-access-9bjb7\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427229 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65b890a-db52-4342-938c-d56538e76b3c-run-httpd\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427260 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-config-data\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.427881 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65b890a-db52-4342-938c-d56538e76b3c-log-httpd\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.428117 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c65b890a-db52-4342-938c-d56538e76b3c-run-httpd\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.434457 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-public-tls-certs\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.435976 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-config-data\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.462426 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-combined-ca-bundle\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.463191 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c65b890a-db52-4342-938c-d56538e76b3c-internal-tls-certs\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.476749 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c65b890a-db52-4342-938c-d56538e76b3c-etc-swift\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.477407 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bjb7\" (UniqueName: \"kubernetes.io/projected/c65b890a-db52-4342-938c-d56538e76b3c-kube-api-access-9bjb7\") pod \"swift-proxy-6568564dcc-mst5g\" (UID: \"c65b890a-db52-4342-938c-d56538e76b3c\") " pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:35 crc kubenswrapper[4796]: I1127 11:57:35.609182 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.163658 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6568564dcc-mst5g"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.695210 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-jcs6h"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.697378 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.705229 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jcs6h"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.758579 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b498531-2aa8-4af7-bbc5-cfeb582a6813-operator-scripts\") pod \"nova-api-db-create-jcs6h\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.758628 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jgkc\" (UniqueName: \"kubernetes.io/projected/0b498531-2aa8-4af7-bbc5-cfeb582a6813-kube-api-access-9jgkc\") pod \"nova-api-db-create-jcs6h\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.798752 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-67mjj"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.800226 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.824406 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.824748 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-central-agent" containerID="cri-o://a1682de5b79877a5fa2df16f1765d759455dad04d13ee69478463532ac899deb" gracePeriod=30 Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.825442 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="proxy-httpd" containerID="cri-o://2be13e35bc18b161ce897b5509527b181d5f746119d5d48ed7ae2c70e25433a3" gracePeriod=30 Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.825518 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-notification-agent" containerID="cri-o://531045a7a9b35b5292ea4d0d2a253bc76988ac519e3a38e9362e097da7a87f2e" gracePeriod=30 Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.825546 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="sg-core" containerID="cri-o://397181c8cad6014c0dcf698f94e281712dc9ca36a9fedc8036933831e64e2977" gracePeriod=30 Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.832767 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.841668 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-67mjj"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.861460 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk6wc\" (UniqueName: \"kubernetes.io/projected/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-kube-api-access-lk6wc\") pod \"nova-cell0-db-create-67mjj\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.861712 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b498531-2aa8-4af7-bbc5-cfeb582a6813-operator-scripts\") pod \"nova-api-db-create-jcs6h\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.861744 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jgkc\" (UniqueName: \"kubernetes.io/projected/0b498531-2aa8-4af7-bbc5-cfeb582a6813-kube-api-access-9jgkc\") pod \"nova-api-db-create-jcs6h\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.861795 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-operator-scripts\") pod \"nova-cell0-db-create-67mjj\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.862930 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b498531-2aa8-4af7-bbc5-cfeb582a6813-operator-scripts\") pod \"nova-api-db-create-jcs6h\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.895554 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jgkc\" (UniqueName: \"kubernetes.io/projected/0b498531-2aa8-4af7-bbc5-cfeb582a6813-kube-api-access-9jgkc\") pod \"nova-api-db-create-jcs6h\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.910018 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e653-account-create-update-4wv78"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.911643 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.913375 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.933440 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e653-account-create-update-4wv78"] Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.963073 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-operator-scripts\") pod \"nova-cell0-db-create-67mjj\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.963133 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7spqr\" (UniqueName: \"kubernetes.io/projected/ceebc602-7a6a-4fc6-baf7-483ca870a28f-kube-api-access-7spqr\") pod \"nova-api-e653-account-create-update-4wv78\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.963249 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ceebc602-7a6a-4fc6-baf7-483ca870a28f-operator-scripts\") pod \"nova-api-e653-account-create-update-4wv78\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.963338 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk6wc\" (UniqueName: \"kubernetes.io/projected/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-kube-api-access-lk6wc\") pod \"nova-cell0-db-create-67mjj\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:36 crc kubenswrapper[4796]: I1127 11:57:36.965713 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-operator-scripts\") pod \"nova-cell0-db-create-67mjj\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:36.998996 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bbj62"] Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.000218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.010435 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk6wc\" (UniqueName: \"kubernetes.io/projected/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-kube-api-access-lk6wc\") pod \"nova-cell0-db-create-67mjj\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.018010 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bbj62"] Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.040038 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.067344 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ceebc602-7a6a-4fc6-baf7-483ca870a28f-operator-scripts\") pod \"nova-api-e653-account-create-update-4wv78\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.067395 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2qmg\" (UniqueName: \"kubernetes.io/projected/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-kube-api-access-c2qmg\") pod \"nova-cell1-db-create-bbj62\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.067518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7spqr\" (UniqueName: \"kubernetes.io/projected/ceebc602-7a6a-4fc6-baf7-483ca870a28f-kube-api-access-7spqr\") pod \"nova-api-e653-account-create-update-4wv78\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.067611 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-operator-scripts\") pod \"nova-cell1-db-create-bbj62\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.068790 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ceebc602-7a6a-4fc6-baf7-483ca870a28f-operator-scripts\") pod \"nova-api-e653-account-create-update-4wv78\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.091167 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7spqr\" (UniqueName: \"kubernetes.io/projected/ceebc602-7a6a-4fc6-baf7-483ca870a28f-kube-api-access-7spqr\") pod \"nova-api-e653-account-create-update-4wv78\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.104548 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-4e96-account-create-update-dmgf7"] Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.106207 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.108387 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.123343 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4e96-account-create-update-dmgf7"] Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.125518 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.171538 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-operator-scripts\") pod \"nova-cell1-db-create-bbj62\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.171624 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/813d270c-4a48-481f-ae6b-763f0f89fb43-operator-scripts\") pod \"nova-cell0-4e96-account-create-update-dmgf7\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.171679 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2qmg\" (UniqueName: \"kubernetes.io/projected/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-kube-api-access-c2qmg\") pod \"nova-cell1-db-create-bbj62\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.172700 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kl2r\" (UniqueName: \"kubernetes.io/projected/813d270c-4a48-481f-ae6b-763f0f89fb43-kube-api-access-8kl2r\") pod \"nova-cell0-4e96-account-create-update-dmgf7\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.173881 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-operator-scripts\") pod \"nova-cell1-db-create-bbj62\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.193102 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2qmg\" (UniqueName: \"kubernetes.io/projected/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-kube-api-access-c2qmg\") pod \"nova-cell1-db-create-bbj62\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.199089 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerID="2be13e35bc18b161ce897b5509527b181d5f746119d5d48ed7ae2c70e25433a3" exitCode=0 Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.199118 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerID="397181c8cad6014c0dcf698f94e281712dc9ca36a9fedc8036933831e64e2977" exitCode=2 Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.199146 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerDied","Data":"2be13e35bc18b161ce897b5509527b181d5f746119d5d48ed7ae2c70e25433a3"} Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.199171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerDied","Data":"397181c8cad6014c0dcf698f94e281712dc9ca36a9fedc8036933831e64e2977"} Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.202176 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6568564dcc-mst5g" event={"ID":"c65b890a-db52-4342-938c-d56538e76b3c","Type":"ContainerStarted","Data":"b9e1902f582f907f5009ebf89b6ba641ce692112c437bf05f96ee0a54cc1eacf"} Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.202210 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6568564dcc-mst5g" event={"ID":"c65b890a-db52-4342-938c-d56538e76b3c","Type":"ContainerStarted","Data":"5aee5df26559eaf7ef407c368883b0b275bee2a7001d05354268e9fe4ca4b2bb"} Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.202220 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6568564dcc-mst5g" event={"ID":"c65b890a-db52-4342-938c-d56538e76b3c","Type":"ContainerStarted","Data":"7fad29a4db3f2c26247f38ffab3a8920a9b146dac4e92488ca9f4d45769e4b6f"} Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.202369 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.202410 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.227562 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6568564dcc-mst5g" podStartSLOduration=2.227535777 podStartE2EDuration="2.227535777s" podCreationTimestamp="2025-11-27 11:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:37.221391293 +0000 UTC m=+1974.739710231" watchObservedRunningTime="2025-11-27 11:57:37.227535777 +0000 UTC m=+1974.745854695" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.261217 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.281567 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/813d270c-4a48-481f-ae6b-763f0f89fb43-operator-scripts\") pod \"nova-cell0-4e96-account-create-update-dmgf7\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.281751 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kl2r\" (UniqueName: \"kubernetes.io/projected/813d270c-4a48-481f-ae6b-763f0f89fb43-kube-api-access-8kl2r\") pod \"nova-cell0-4e96-account-create-update-dmgf7\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.284257 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/813d270c-4a48-481f-ae6b-763f0f89fb43-operator-scripts\") pod \"nova-cell0-4e96-account-create-update-dmgf7\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.310000 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5a51-account-create-update-s8rhc"] Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.310697 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kl2r\" (UniqueName: \"kubernetes.io/projected/813d270c-4a48-481f-ae6b-763f0f89fb43-kube-api-access-8kl2r\") pod \"nova-cell0-4e96-account-create-update-dmgf7\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.311500 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.315914 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.342537 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5a51-account-create-update-s8rhc"] Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.360854 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.384268 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8af3632d-426e-4004-b787-66529242255b-operator-scripts\") pod \"nova-cell1-5a51-account-create-update-s8rhc\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.384350 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7jgh\" (UniqueName: \"kubernetes.io/projected/8af3632d-426e-4004-b787-66529242255b-kube-api-access-b7jgh\") pod \"nova-cell1-5a51-account-create-update-s8rhc\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.453512 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.486361 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8af3632d-426e-4004-b787-66529242255b-operator-scripts\") pod \"nova-cell1-5a51-account-create-update-s8rhc\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.486428 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7jgh\" (UniqueName: \"kubernetes.io/projected/8af3632d-426e-4004-b787-66529242255b-kube-api-access-b7jgh\") pod \"nova-cell1-5a51-account-create-update-s8rhc\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.487411 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8af3632d-426e-4004-b787-66529242255b-operator-scripts\") pod \"nova-cell1-5a51-account-create-update-s8rhc\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.518850 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7jgh\" (UniqueName: \"kubernetes.io/projected/8af3632d-426e-4004-b787-66529242255b-kube-api-access-b7jgh\") pod \"nova-cell1-5a51-account-create-update-s8rhc\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:37 crc kubenswrapper[4796]: I1127 11:57:37.658439 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:38 crc kubenswrapper[4796]: I1127 11:57:38.222854 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerID="a1682de5b79877a5fa2df16f1765d759455dad04d13ee69478463532ac899deb" exitCode=0 Nov 27 11:57:38 crc kubenswrapper[4796]: I1127 11:57:38.222927 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerDied","Data":"a1682de5b79877a5fa2df16f1765d759455dad04d13ee69478463532ac899deb"} Nov 27 11:57:39 crc kubenswrapper[4796]: I1127 11:57:39.235238 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerID="531045a7a9b35b5292ea4d0d2a253bc76988ac519e3a38e9362e097da7a87f2e" exitCode=0 Nov 27 11:57:39 crc kubenswrapper[4796]: I1127 11:57:39.235318 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerDied","Data":"531045a7a9b35b5292ea4d0d2a253bc76988ac519e3a38e9362e097da7a87f2e"} Nov 27 11:57:39 crc kubenswrapper[4796]: I1127 11:57:39.740424 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 27 11:57:40 crc kubenswrapper[4796]: I1127 11:57:40.415175 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:40 crc kubenswrapper[4796]: I1127 11:57:40.417558 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d4b54c868-49fn9" Nov 27 11:57:41 crc kubenswrapper[4796]: I1127 11:57:41.569791 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:57:41 crc kubenswrapper[4796]: E1127 11:57:41.570325 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.606424 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806553 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-config-data\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806644 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-log-httpd\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806685 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-sg-core-conf-yaml\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806749 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt4qx\" (UniqueName: \"kubernetes.io/projected/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-kube-api-access-jt4qx\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806781 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-combined-ca-bundle\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806895 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-run-httpd\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.806967 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-scripts\") pod \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\" (UID: \"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac\") " Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.808704 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.808823 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.813072 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-scripts" (OuterVolumeSpecName: "scripts") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.813984 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-kube-api-access-jt4qx" (OuterVolumeSpecName: "kube-api-access-jt4qx") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "kube-api-access-jt4qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.853627 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.895086 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.908925 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.908964 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.908976 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.908988 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt4qx\" (UniqueName: \"kubernetes.io/projected/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-kube-api-access-jt4qx\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.908999 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.909011 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:43 crc kubenswrapper[4796]: I1127 11:57:43.919567 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-config-data" (OuterVolumeSpecName: "config-data") pod "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" (UID: "e9482d56-62c4-41e8-b9ee-d5e9115cf8ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.011485 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:44 crc kubenswrapper[4796]: W1127 11:57:44.198666 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3a975cf_a1d9_4cc3_ad76_4e60392c6fc3.slice/crio-507b67d4eada673dbac0ffed42e83c109566e281dab450f75c0996f9f85fb0ae WatchSource:0}: Error finding container 507b67d4eada673dbac0ffed42e83c109566e281dab450f75c0996f9f85fb0ae: Status 404 returned error can't find the container with id 507b67d4eada673dbac0ffed42e83c109566e281dab450f75c0996f9f85fb0ae Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.202052 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-67mjj"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.233841 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jcs6h"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.242857 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.247354 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4e96-account-create-update-dmgf7"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.256909 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e653-account-create-update-4wv78"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.267869 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bbj62"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.273643 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5a51-account-create-update-s8rhc"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.281471 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.309764 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.312400 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jcs6h" event={"ID":"0b498531-2aa8-4af7-bbc5-cfeb582a6813","Type":"ContainerStarted","Data":"add9eafccbd705e0aefe64f6531468f0e2c50618016c4a7649b5e7de33d9383f"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.318451 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e653-account-create-update-4wv78" event={"ID":"ceebc602-7a6a-4fc6-baf7-483ca870a28f","Type":"ContainerStarted","Data":"42d61f0ade7452fcc8c8a799882c7c7cf593424fb800e7b86c547aa656fc1873"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.319803 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" event={"ID":"8af3632d-426e-4004-b787-66529242255b","Type":"ContainerStarted","Data":"91fc7afd199b5a79039dfec30dd73758d806d5ff35340500336c4fb2b6a29040"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.331792 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9482d56-62c4-41e8-b9ee-d5e9115cf8ac","Type":"ContainerDied","Data":"7dc4eda8d03be1993e2e1653aff83f4a9264945ac27826c7259d7473abf2997e"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.332170 4796 scope.go:117] "RemoveContainer" containerID="2be13e35bc18b161ce897b5509527b181d5f746119d5d48ed7ae2c70e25433a3" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.332804 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.349364 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" event={"ID":"813d270c-4a48-481f-ae6b-763f0f89fb43","Type":"ContainerStarted","Data":"2bcd4dce34e67b551bf0301ca6114329fa9faa38d23987fffa40b41de8f5476c"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.356060 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-67mjj" event={"ID":"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3","Type":"ContainerStarted","Data":"507b67d4eada673dbac0ffed42e83c109566e281dab450f75c0996f9f85fb0ae"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.357561 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bbj62" event={"ID":"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1","Type":"ContainerStarted","Data":"b5b9950ef1da203677cd8f12e5c4503f32b58729927e6b7641866e87915d45eb"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.362054 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c54eb17e-bce8-4e91-b9f2-766040bcfc89","Type":"ContainerStarted","Data":"0cb2cdfd46668f8acf86b264ecdf74a878452ee46e557baf5cb460a79307f45f"} Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.385068 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.8220597 podStartE2EDuration="13.385050757s" podCreationTimestamp="2025-11-27 11:57:31 +0000 UTC" firstStartedPulling="2025-11-27 11:57:32.744677748 +0000 UTC m=+1970.262996666" lastFinishedPulling="2025-11-27 11:57:43.307668795 +0000 UTC m=+1980.825987723" observedRunningTime="2025-11-27 11:57:44.380741852 +0000 UTC m=+1981.899060770" watchObservedRunningTime="2025-11-27 11:57:44.385050757 +0000 UTC m=+1981.903369675" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.401496 4796 scope.go:117] "RemoveContainer" containerID="397181c8cad6014c0dcf698f94e281712dc9ca36a9fedc8036933831e64e2977" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.430439 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.446222 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.458413 4796 scope.go:117] "RemoveContainer" containerID="531045a7a9b35b5292ea4d0d2a253bc76988ac519e3a38e9362e097da7a87f2e" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.466916 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:44 crc kubenswrapper[4796]: E1127 11:57:44.467488 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-central-agent" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467528 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-central-agent" Nov 27 11:57:44 crc kubenswrapper[4796]: E1127 11:57:44.467541 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="proxy-httpd" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467547 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="proxy-httpd" Nov 27 11:57:44 crc kubenswrapper[4796]: E1127 11:57:44.467582 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="sg-core" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467588 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="sg-core" Nov 27 11:57:44 crc kubenswrapper[4796]: E1127 11:57:44.467610 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-notification-agent" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467617 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-notification-agent" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467913 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="sg-core" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467945 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="proxy-httpd" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467961 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-central-agent" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.467990 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" containerName="ceilometer-notification-agent" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.470169 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.477612 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.478201 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.478625 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.503034 4796 scope.go:117] "RemoveContainer" containerID="a1682de5b79877a5fa2df16f1765d759455dad04d13ee69478463532ac899deb" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.524502 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.524911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-log-httpd\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.525051 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-run-httpd\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.525132 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.525202 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-scripts\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.525312 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjjwt\" (UniqueName: \"kubernetes.io/projected/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-kube-api-access-hjjwt\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.525403 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-config-data\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.626866 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.628404 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-log-httpd\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.628531 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-run-httpd\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.628640 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.628742 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-scripts\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.628929 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjjwt\" (UniqueName: \"kubernetes.io/projected/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-kube-api-access-hjjwt\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.629050 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-config-data\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.628984 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-log-httpd\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.629057 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-run-httpd\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.647184 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-scripts\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.647230 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.647944 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.650918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-config-data\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.651182 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjjwt\" (UniqueName: \"kubernetes.io/projected/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-kube-api-access-hjjwt\") pod \"ceilometer-0\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " pod="openstack/ceilometer-0" Nov 27 11:57:44 crc kubenswrapper[4796]: I1127 11:57:44.812413 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.177976 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.372645 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" event={"ID":"8af3632d-426e-4004-b787-66529242255b","Type":"ContainerStarted","Data":"e3e3526d04f419488f28c8c41e468a5086ac8c7dbaf8e74b572a62d692ec758b"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.375599 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" event={"ID":"813d270c-4a48-481f-ae6b-763f0f89fb43","Type":"ContainerStarted","Data":"c271eeeead555e12c731437950972e58c38d25b0fa872ef41340751282bcb26b"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.378389 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-67mjj" event={"ID":"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3","Type":"ContainerStarted","Data":"44ea6ff19a93b29f6f3647cbb3c4692880f04249bbcfae1cd96727d2cb76b0b3"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.380160 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bbj62" event={"ID":"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1","Type":"ContainerStarted","Data":"6ea459223ec6ce9b06b2dd9da5fcadb6836a5c10a7142318314ab8cf5707671b"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.382171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerStarted","Data":"e42563294f5e7b145827fb06ec64a10ea9c1001e34612681b0191ea17415d561"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.383638 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jcs6h" event={"ID":"0b498531-2aa8-4af7-bbc5-cfeb582a6813","Type":"ContainerStarted","Data":"5794db6ea70b2cc1983c289178a1e8ff0817f382d265082865628ca8284b188f"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.386327 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e653-account-create-update-4wv78" event={"ID":"ceebc602-7a6a-4fc6-baf7-483ca870a28f","Type":"ContainerStarted","Data":"d51dbf066d77c2ee6d11edcb605a311da42714d36fd4aff7c9918867e96fc70d"} Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.399453 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" podStartSLOduration=8.399435308 podStartE2EDuration="8.399435308s" podCreationTimestamp="2025-11-27 11:57:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:45.391502576 +0000 UTC m=+1982.909821494" watchObservedRunningTime="2025-11-27 11:57:45.399435308 +0000 UTC m=+1982.917754226" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.425832 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-jcs6h" podStartSLOduration=9.425814571 podStartE2EDuration="9.425814571s" podCreationTimestamp="2025-11-27 11:57:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:45.41453765 +0000 UTC m=+1982.932856588" watchObservedRunningTime="2025-11-27 11:57:45.425814571 +0000 UTC m=+1982.944133489" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.442525 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-bbj62" podStartSLOduration=9.442504286 podStartE2EDuration="9.442504286s" podCreationTimestamp="2025-11-27 11:57:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:45.428900023 +0000 UTC m=+1982.947218941" watchObservedRunningTime="2025-11-27 11:57:45.442504286 +0000 UTC m=+1982.960823204" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.446378 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" podStartSLOduration=8.44636332 podStartE2EDuration="8.44636332s" podCreationTimestamp="2025-11-27 11:57:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:45.443523954 +0000 UTC m=+1982.961842872" watchObservedRunningTime="2025-11-27 11:57:45.44636332 +0000 UTC m=+1982.964682238" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.462412 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-e653-account-create-update-4wv78" podStartSLOduration=9.462388327 podStartE2EDuration="9.462388327s" podCreationTimestamp="2025-11-27 11:57:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:45.45877352 +0000 UTC m=+1982.977092438" watchObservedRunningTime="2025-11-27 11:57:45.462388327 +0000 UTC m=+1982.980707245" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.590553 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9482d56-62c4-41e8-b9ee-d5e9115cf8ac" path="/var/lib/kubelet/pods/e9482d56-62c4-41e8-b9ee-d5e9115cf8ac/volumes" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.615322 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:45 crc kubenswrapper[4796]: I1127 11:57:45.630073 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6568564dcc-mst5g" Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.395313 4796 generic.go:334] "Generic (PLEG): container finished" podID="0b498531-2aa8-4af7-bbc5-cfeb582a6813" containerID="5794db6ea70b2cc1983c289178a1e8ff0817f382d265082865628ca8284b188f" exitCode=0 Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.395806 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jcs6h" event={"ID":"0b498531-2aa8-4af7-bbc5-cfeb582a6813","Type":"ContainerDied","Data":"5794db6ea70b2cc1983c289178a1e8ff0817f382d265082865628ca8284b188f"} Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.399852 4796 generic.go:334] "Generic (PLEG): container finished" podID="ceebc602-7a6a-4fc6-baf7-483ca870a28f" containerID="d51dbf066d77c2ee6d11edcb605a311da42714d36fd4aff7c9918867e96fc70d" exitCode=0 Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.399904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e653-account-create-update-4wv78" event={"ID":"ceebc602-7a6a-4fc6-baf7-483ca870a28f","Type":"ContainerDied","Data":"d51dbf066d77c2ee6d11edcb605a311da42714d36fd4aff7c9918867e96fc70d"} Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.401443 4796 generic.go:334] "Generic (PLEG): container finished" podID="8af3632d-426e-4004-b787-66529242255b" containerID="e3e3526d04f419488f28c8c41e468a5086ac8c7dbaf8e74b572a62d692ec758b" exitCode=0 Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.401484 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" event={"ID":"8af3632d-426e-4004-b787-66529242255b","Type":"ContainerDied","Data":"e3e3526d04f419488f28c8c41e468a5086ac8c7dbaf8e74b572a62d692ec758b"} Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.402697 4796 generic.go:334] "Generic (PLEG): container finished" podID="813d270c-4a48-481f-ae6b-763f0f89fb43" containerID="c271eeeead555e12c731437950972e58c38d25b0fa872ef41340751282bcb26b" exitCode=0 Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.402734 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" event={"ID":"813d270c-4a48-481f-ae6b-763f0f89fb43","Type":"ContainerDied","Data":"c271eeeead555e12c731437950972e58c38d25b0fa872ef41340751282bcb26b"} Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.404168 4796 generic.go:334] "Generic (PLEG): container finished" podID="d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" containerID="44ea6ff19a93b29f6f3647cbb3c4692880f04249bbcfae1cd96727d2cb76b0b3" exitCode=0 Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.404253 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-67mjj" event={"ID":"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3","Type":"ContainerDied","Data":"44ea6ff19a93b29f6f3647cbb3c4692880f04249bbcfae1cd96727d2cb76b0b3"} Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.410427 4796 generic.go:334] "Generic (PLEG): container finished" podID="bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" containerID="6ea459223ec6ce9b06b2dd9da5fcadb6836a5c10a7142318314ab8cf5707671b" exitCode=0 Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.410533 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bbj62" event={"ID":"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1","Type":"ContainerDied","Data":"6ea459223ec6ce9b06b2dd9da5fcadb6836a5c10a7142318314ab8cf5707671b"} Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.777650 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.877655 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk6wc\" (UniqueName: \"kubernetes.io/projected/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-kube-api-access-lk6wc\") pod \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.878003 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-operator-scripts\") pod \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\" (UID: \"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3\") " Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.878732 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" (UID: "d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.884448 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-kube-api-access-lk6wc" (OuterVolumeSpecName: "kube-api-access-lk6wc") pod "d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" (UID: "d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3"). InnerVolumeSpecName "kube-api-access-lk6wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.980475 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk6wc\" (UniqueName: \"kubernetes.io/projected/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-kube-api-access-lk6wc\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:46 crc kubenswrapper[4796]: I1127 11:57:46.980517 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.313387 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.422142 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-67mjj" Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.422141 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-67mjj" event={"ID":"d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3","Type":"ContainerDied","Data":"507b67d4eada673dbac0ffed42e83c109566e281dab450f75c0996f9f85fb0ae"} Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.422275 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="507b67d4eada673dbac0ffed42e83c109566e281dab450f75c0996f9f85fb0ae" Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.424109 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerStarted","Data":"88121c6cae30fd137788d814fb2e15e2d298a0bd120cc42249dabfaf03062d0e"} Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.424144 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerStarted","Data":"69eb6ccc1e17b4f283f653793de69ab9620fb5164617ca835b2fda897c525557"} Nov 27 11:57:47 crc kubenswrapper[4796]: I1127 11:57:47.914814 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.112363 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b498531-2aa8-4af7-bbc5-cfeb582a6813-operator-scripts\") pod \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.112614 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jgkc\" (UniqueName: \"kubernetes.io/projected/0b498531-2aa8-4af7-bbc5-cfeb582a6813-kube-api-access-9jgkc\") pod \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\" (UID: \"0b498531-2aa8-4af7-bbc5-cfeb582a6813\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.114734 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b498531-2aa8-4af7-bbc5-cfeb582a6813-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0b498531-2aa8-4af7-bbc5-cfeb582a6813" (UID: "0b498531-2aa8-4af7-bbc5-cfeb582a6813"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.119154 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b498531-2aa8-4af7-bbc5-cfeb582a6813-kube-api-access-9jgkc" (OuterVolumeSpecName: "kube-api-access-9jgkc") pod "0b498531-2aa8-4af7-bbc5-cfeb582a6813" (UID: "0b498531-2aa8-4af7-bbc5-cfeb582a6813"). InnerVolumeSpecName "kube-api-access-9jgkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.130102 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.189610 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.195802 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.202819 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.218652 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b498531-2aa8-4af7-bbc5-cfeb582a6813-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.218761 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jgkc\" (UniqueName: \"kubernetes.io/projected/0b498531-2aa8-4af7-bbc5-cfeb582a6813-kube-api-access-9jgkc\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.320852 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-operator-scripts\") pod \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.320889 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kl2r\" (UniqueName: \"kubernetes.io/projected/813d270c-4a48-481f-ae6b-763f0f89fb43-kube-api-access-8kl2r\") pod \"813d270c-4a48-481f-ae6b-763f0f89fb43\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.320949 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/813d270c-4a48-481f-ae6b-763f0f89fb43-operator-scripts\") pod \"813d270c-4a48-481f-ae6b-763f0f89fb43\" (UID: \"813d270c-4a48-481f-ae6b-763f0f89fb43\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.320967 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2qmg\" (UniqueName: \"kubernetes.io/projected/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-kube-api-access-c2qmg\") pod \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\" (UID: \"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.321099 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7jgh\" (UniqueName: \"kubernetes.io/projected/8af3632d-426e-4004-b787-66529242255b-kube-api-access-b7jgh\") pod \"8af3632d-426e-4004-b787-66529242255b\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.321128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8af3632d-426e-4004-b787-66529242255b-operator-scripts\") pod \"8af3632d-426e-4004-b787-66529242255b\" (UID: \"8af3632d-426e-4004-b787-66529242255b\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.321163 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7spqr\" (UniqueName: \"kubernetes.io/projected/ceebc602-7a6a-4fc6-baf7-483ca870a28f-kube-api-access-7spqr\") pod \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.321186 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ceebc602-7a6a-4fc6-baf7-483ca870a28f-operator-scripts\") pod \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\" (UID: \"ceebc602-7a6a-4fc6-baf7-483ca870a28f\") " Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.321922 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceebc602-7a6a-4fc6-baf7-483ca870a28f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ceebc602-7a6a-4fc6-baf7-483ca870a28f" (UID: "ceebc602-7a6a-4fc6-baf7-483ca870a28f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.322651 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" (UID: "bcb76cd8-08df-45f1-ac77-6b8d326ed4d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.323081 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/813d270c-4a48-481f-ae6b-763f0f89fb43-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "813d270c-4a48-481f-ae6b-763f0f89fb43" (UID: "813d270c-4a48-481f-ae6b-763f0f89fb43"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.323447 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8af3632d-426e-4004-b787-66529242255b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8af3632d-426e-4004-b787-66529242255b" (UID: "8af3632d-426e-4004-b787-66529242255b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.327220 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceebc602-7a6a-4fc6-baf7-483ca870a28f-kube-api-access-7spqr" (OuterVolumeSpecName: "kube-api-access-7spqr") pod "ceebc602-7a6a-4fc6-baf7-483ca870a28f" (UID: "ceebc602-7a6a-4fc6-baf7-483ca870a28f"). InnerVolumeSpecName "kube-api-access-7spqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.327444 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/813d270c-4a48-481f-ae6b-763f0f89fb43-kube-api-access-8kl2r" (OuterVolumeSpecName: "kube-api-access-8kl2r") pod "813d270c-4a48-481f-ae6b-763f0f89fb43" (UID: "813d270c-4a48-481f-ae6b-763f0f89fb43"). InnerVolumeSpecName "kube-api-access-8kl2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.327578 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-kube-api-access-c2qmg" (OuterVolumeSpecName: "kube-api-access-c2qmg") pod "bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" (UID: "bcb76cd8-08df-45f1-ac77-6b8d326ed4d1"). InnerVolumeSpecName "kube-api-access-c2qmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.329836 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8af3632d-426e-4004-b787-66529242255b-kube-api-access-b7jgh" (OuterVolumeSpecName: "kube-api-access-b7jgh") pod "8af3632d-426e-4004-b787-66529242255b" (UID: "8af3632d-426e-4004-b787-66529242255b"). InnerVolumeSpecName "kube-api-access-b7jgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424401 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7jgh\" (UniqueName: \"kubernetes.io/projected/8af3632d-426e-4004-b787-66529242255b-kube-api-access-b7jgh\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424430 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8af3632d-426e-4004-b787-66529242255b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424441 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7spqr\" (UniqueName: \"kubernetes.io/projected/ceebc602-7a6a-4fc6-baf7-483ca870a28f-kube-api-access-7spqr\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424450 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ceebc602-7a6a-4fc6-baf7-483ca870a28f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424458 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424466 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kl2r\" (UniqueName: \"kubernetes.io/projected/813d270c-4a48-481f-ae6b-763f0f89fb43-kube-api-access-8kl2r\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424475 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/813d270c-4a48-481f-ae6b-763f0f89fb43-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.424485 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2qmg\" (UniqueName: \"kubernetes.io/projected/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1-kube-api-access-c2qmg\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.457962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jcs6h" event={"ID":"0b498531-2aa8-4af7-bbc5-cfeb582a6813","Type":"ContainerDied","Data":"add9eafccbd705e0aefe64f6531468f0e2c50618016c4a7649b5e7de33d9383f"} Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.458020 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="add9eafccbd705e0aefe64f6531468f0e2c50618016c4a7649b5e7de33d9383f" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.458158 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jcs6h" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.460746 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e653-account-create-update-4wv78" event={"ID":"ceebc602-7a6a-4fc6-baf7-483ca870a28f","Type":"ContainerDied","Data":"42d61f0ade7452fcc8c8a799882c7c7cf593424fb800e7b86c547aa656fc1873"} Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.460787 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42d61f0ade7452fcc8c8a799882c7c7cf593424fb800e7b86c547aa656fc1873" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.460889 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e653-account-create-update-4wv78" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.469208 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" event={"ID":"8af3632d-426e-4004-b787-66529242255b","Type":"ContainerDied","Data":"91fc7afd199b5a79039dfec30dd73758d806d5ff35340500336c4fb2b6a29040"} Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.469257 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91fc7afd199b5a79039dfec30dd73758d806d5ff35340500336c4fb2b6a29040" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.469336 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a51-account-create-update-s8rhc" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.488040 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.488403 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4e96-account-create-update-dmgf7" event={"ID":"813d270c-4a48-481f-ae6b-763f0f89fb43","Type":"ContainerDied","Data":"2bcd4dce34e67b551bf0301ca6114329fa9faa38d23987fffa40b41de8f5476c"} Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.488453 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bcd4dce34e67b551bf0301ca6114329fa9faa38d23987fffa40b41de8f5476c" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.493076 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bbj62" event={"ID":"bcb76cd8-08df-45f1-ac77-6b8d326ed4d1","Type":"ContainerDied","Data":"b5b9950ef1da203677cd8f12e5c4503f32b58729927e6b7641866e87915d45eb"} Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.493115 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5b9950ef1da203677cd8f12e5c4503f32b58729927e6b7641866e87915d45eb" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.493191 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bbj62" Nov 27 11:57:48 crc kubenswrapper[4796]: I1127 11:57:48.508250 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerStarted","Data":"967961baae2e08bacffd81376bdbc6f6a9447cddc3481c1f4f37c47db835d6ae"} Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.542674 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerStarted","Data":"0f2e3c2b77e906cfec7058b4a27e9ee75587f66a0a0552f5af67cf5c9a0460e5"} Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.543194 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-central-agent" containerID="cri-o://69eb6ccc1e17b4f283f653793de69ab9620fb5164617ca835b2fda897c525557" gracePeriod=30 Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.543292 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.543437 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="proxy-httpd" containerID="cri-o://0f2e3c2b77e906cfec7058b4a27e9ee75587f66a0a0552f5af67cf5c9a0460e5" gracePeriod=30 Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.543506 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="sg-core" containerID="cri-o://967961baae2e08bacffd81376bdbc6f6a9447cddc3481c1f4f37c47db835d6ae" gracePeriod=30 Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.543565 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-notification-agent" containerID="cri-o://88121c6cae30fd137788d814fb2e15e2d298a0bd120cc42249dabfaf03062d0e" gracePeriod=30 Nov 27 11:57:50 crc kubenswrapper[4796]: I1127 11:57:50.561622 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3004654540000002 podStartE2EDuration="6.561605729s" podCreationTimestamp="2025-11-27 11:57:44 +0000 UTC" firstStartedPulling="2025-11-27 11:57:45.201521877 +0000 UTC m=+1982.719840795" lastFinishedPulling="2025-11-27 11:57:49.462662152 +0000 UTC m=+1986.980981070" observedRunningTime="2025-11-27 11:57:50.560646573 +0000 UTC m=+1988.078965491" watchObservedRunningTime="2025-11-27 11:57:50.561605729 +0000 UTC m=+1988.079924647" Nov 27 11:57:51 crc kubenswrapper[4796]: I1127 11:57:51.553096 4796 generic.go:334] "Generic (PLEG): container finished" podID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerID="0f2e3c2b77e906cfec7058b4a27e9ee75587f66a0a0552f5af67cf5c9a0460e5" exitCode=0 Nov 27 11:57:51 crc kubenswrapper[4796]: I1127 11:57:51.553626 4796 generic.go:334] "Generic (PLEG): container finished" podID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerID="967961baae2e08bacffd81376bdbc6f6a9447cddc3481c1f4f37c47db835d6ae" exitCode=2 Nov 27 11:57:51 crc kubenswrapper[4796]: I1127 11:57:51.553639 4796 generic.go:334] "Generic (PLEG): container finished" podID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerID="88121c6cae30fd137788d814fb2e15e2d298a0bd120cc42249dabfaf03062d0e" exitCode=0 Nov 27 11:57:51 crc kubenswrapper[4796]: I1127 11:57:51.553660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerDied","Data":"0f2e3c2b77e906cfec7058b4a27e9ee75587f66a0a0552f5af67cf5c9a0460e5"} Nov 27 11:57:51 crc kubenswrapper[4796]: I1127 11:57:51.553686 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerDied","Data":"967961baae2e08bacffd81376bdbc6f6a9447cddc3481c1f4f37c47db835d6ae"} Nov 27 11:57:51 crc kubenswrapper[4796]: I1127 11:57:51.553698 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerDied","Data":"88121c6cae30fd137788d814fb2e15e2d298a0bd120cc42249dabfaf03062d0e"} Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.391912 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-whbbq"] Nov 27 11:57:52 crc kubenswrapper[4796]: E1127 11:57:52.392261 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceebc602-7a6a-4fc6-baf7-483ca870a28f" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392290 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceebc602-7a6a-4fc6-baf7-483ca870a28f" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: E1127 11:57:52.392308 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392315 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: E1127 11:57:52.392331 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392338 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: E1127 11:57:52.392357 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="813d270c-4a48-481f-ae6b-763f0f89fb43" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392363 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="813d270c-4a48-481f-ae6b-763f0f89fb43" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: E1127 11:57:52.392369 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af3632d-426e-4004-b787-66529242255b" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392375 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af3632d-426e-4004-b787-66529242255b" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: E1127 11:57:52.392394 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b498531-2aa8-4af7-bbc5-cfeb582a6813" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392401 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b498531-2aa8-4af7-bbc5-cfeb582a6813" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392548 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392563 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392574 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b498531-2aa8-4af7-bbc5-cfeb582a6813" containerName="mariadb-database-create" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392584 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="813d270c-4a48-481f-ae6b-763f0f89fb43" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392592 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceebc602-7a6a-4fc6-baf7-483ca870a28f" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.392599 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af3632d-426e-4004-b787-66529242255b" containerName="mariadb-account-create-update" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.393116 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.395228 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.395236 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.398400 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4htnn" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.450708 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-whbbq"] Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.498724 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.498822 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-scripts\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.498857 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-config-data\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.498982 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgf82\" (UniqueName: \"kubernetes.io/projected/6524424e-afc2-443c-beac-c0503ac10ddf-kube-api-access-kgf82\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.602609 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.602680 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-scripts\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.602705 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-config-data\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.602796 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgf82\" (UniqueName: \"kubernetes.io/projected/6524424e-afc2-443c-beac-c0503ac10ddf-kube-api-access-kgf82\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.608662 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-config-data\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.609719 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.611447 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-scripts\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.621469 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgf82\" (UniqueName: \"kubernetes.io/projected/6524424e-afc2-443c-beac-c0503ac10ddf-kube-api-access-kgf82\") pod \"nova-cell0-conductor-db-sync-whbbq\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:52 crc kubenswrapper[4796]: I1127 11:57:52.785479 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.239731 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-658b556c84-2wb2w"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.241408 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.244810 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-24wlq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.244878 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.245171 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.248339 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-whbbq"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.262119 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-658b556c84-2wb2w"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.314848 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data-custom\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.315012 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-combined-ca-bundle\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.315050 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.315075 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l7nb\" (UniqueName: \"kubernetes.io/projected/354047e3-61d2-4505-93f6-86592d3f62a7-kube-api-access-4l7nb\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.417341 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-combined-ca-bundle\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.417386 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.417406 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l7nb\" (UniqueName: \"kubernetes.io/projected/354047e3-61d2-4505-93f6-86592d3f62a7-kube-api-access-4l7nb\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.417517 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data-custom\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.425027 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.425079 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data-custom\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.431349 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-pkmhj"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.433128 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.443170 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-combined-ca-bundle\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.449045 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l7nb\" (UniqueName: \"kubernetes.io/projected/354047e3-61d2-4505-93f6-86592d3f62a7-kube-api-access-4l7nb\") pod \"heat-engine-658b556c84-2wb2w\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.460103 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-f94c75f96-tmxbc"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.461330 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.468628 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.481757 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-pkmhj"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.512084 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f94c75f96-tmxbc"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522040 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dt4n\" (UniqueName: \"kubernetes.io/projected/8cdbd3cc-dbee-4abb-8e34-c910554b5142-kube-api-access-9dt4n\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522094 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzs2r\" (UniqueName: \"kubernetes.io/projected/38cfea73-e382-40ee-a600-7f0c340b9e89-kube-api-access-bzs2r\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522116 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522151 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data-custom\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522165 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522180 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-config\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522203 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522234 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522297 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.522323 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-combined-ca-bundle\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.535879 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5d6d658794-wxmmq"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.537127 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.539645 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.580190 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:57:53 crc kubenswrapper[4796]: E1127 11:57:53.580712 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.582257 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5d6d658794-wxmmq"] Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.582326 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-whbbq" event={"ID":"6524424e-afc2-443c-beac-c0503ac10ddf","Type":"ContainerStarted","Data":"c11ca53fce62e703be022b1b74a6bfff1a7626a983ddb45edbe457cb02313231"} Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625389 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data-custom\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625437 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625462 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-config\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625502 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625548 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625597 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-combined-ca-bundle\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625704 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dt4n\" (UniqueName: \"kubernetes.io/projected/8cdbd3cc-dbee-4abb-8e34-c910554b5142-kube-api-access-9dt4n\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625744 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzs2r\" (UniqueName: \"kubernetes.io/projected/38cfea73-e382-40ee-a600-7f0c340b9e89-kube-api-access-bzs2r\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.625784 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.626868 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.627065 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.628970 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-config\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.629044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.629227 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.629678 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.633471 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.633627 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-combined-ca-bundle\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.638052 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data-custom\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.658455 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzs2r\" (UniqueName: \"kubernetes.io/projected/38cfea73-e382-40ee-a600-7f0c340b9e89-kube-api-access-bzs2r\") pod \"heat-api-f94c75f96-tmxbc\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.662968 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dt4n\" (UniqueName: \"kubernetes.io/projected/8cdbd3cc-dbee-4abb-8e34-c910554b5142-kube-api-access-9dt4n\") pod \"dnsmasq-dns-688b9f5b49-pkmhj\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.729004 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.729086 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-combined-ca-bundle\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.729151 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8w4l\" (UniqueName: \"kubernetes.io/projected/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-kube-api-access-h8w4l\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.729181 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data-custom\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.768357 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.779237 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.831812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.831875 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-combined-ca-bundle\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.831916 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8w4l\" (UniqueName: \"kubernetes.io/projected/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-kube-api-access-h8w4l\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.831936 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data-custom\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.855694 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-combined-ca-bundle\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.864221 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.865157 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data-custom\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.881216 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8w4l\" (UniqueName: \"kubernetes.io/projected/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-kube-api-access-h8w4l\") pod \"heat-cfnapi-5d6d658794-wxmmq\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:53 crc kubenswrapper[4796]: I1127 11:57:53.970103 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-658b556c84-2wb2w"] Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.157700 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.297354 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f94c75f96-tmxbc"] Nov 27 11:57:54 crc kubenswrapper[4796]: W1127 11:57:54.314033 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38cfea73_e382_40ee_a600_7f0c340b9e89.slice/crio-7abd9a975dbcacddab6b30e3a5ce60ccccb3783f25852eccc373cf3132a22f59 WatchSource:0}: Error finding container 7abd9a975dbcacddab6b30e3a5ce60ccccb3783f25852eccc373cf3132a22f59: Status 404 returned error can't find the container with id 7abd9a975dbcacddab6b30e3a5ce60ccccb3783f25852eccc373cf3132a22f59 Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.584834 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-pkmhj"] Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.610468 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-658b556c84-2wb2w" event={"ID":"354047e3-61d2-4505-93f6-86592d3f62a7","Type":"ContainerStarted","Data":"cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21"} Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.610800 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-658b556c84-2wb2w" event={"ID":"354047e3-61d2-4505-93f6-86592d3f62a7","Type":"ContainerStarted","Data":"e18339c50968a657c05bd92131a3e2a2a1fa9eb2d46de45758659310eb93339f"} Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.612281 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.616979 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f94c75f96-tmxbc" event={"ID":"38cfea73-e382-40ee-a600-7f0c340b9e89","Type":"ContainerStarted","Data":"7abd9a975dbcacddab6b30e3a5ce60ccccb3783f25852eccc373cf3132a22f59"} Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.634374 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-658b556c84-2wb2w" podStartSLOduration=1.634353066 podStartE2EDuration="1.634353066s" podCreationTimestamp="2025-11-27 11:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:54.631051449 +0000 UTC m=+1992.149370367" watchObservedRunningTime="2025-11-27 11:57:54.634353066 +0000 UTC m=+1992.152671984" Nov 27 11:57:54 crc kubenswrapper[4796]: I1127 11:57:54.829800 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5d6d658794-wxmmq"] Nov 27 11:57:55 crc kubenswrapper[4796]: I1127 11:57:55.632210 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" event={"ID":"74b266b4-8a6e-4b8f-91c8-de16da0f76e8","Type":"ContainerStarted","Data":"530c5c14306c324f15ffa0dde4472e322e16e46bae7cd304c19a0559b19803ab"} Nov 27 11:57:55 crc kubenswrapper[4796]: I1127 11:57:55.634612 4796 generic.go:334] "Generic (PLEG): container finished" podID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerID="fef71d3d94aab373167ea9ac4c5d3279393a5dd18601243f21ec3aa4741d5f29" exitCode=0 Nov 27 11:57:55 crc kubenswrapper[4796]: I1127 11:57:55.634764 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" event={"ID":"8cdbd3cc-dbee-4abb-8e34-c910554b5142","Type":"ContainerDied","Data":"fef71d3d94aab373167ea9ac4c5d3279393a5dd18601243f21ec3aa4741d5f29"} Nov 27 11:57:55 crc kubenswrapper[4796]: I1127 11:57:55.634811 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" event={"ID":"8cdbd3cc-dbee-4abb-8e34-c910554b5142","Type":"ContainerStarted","Data":"99bdb254076ef589d2b316193793ce17d4368950acd201c2b5330a039ed9b306"} Nov 27 11:57:56 crc kubenswrapper[4796]: I1127 11:57:56.650653 4796 generic.go:334] "Generic (PLEG): container finished" podID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerID="69eb6ccc1e17b4f283f653793de69ab9620fb5164617ca835b2fda897c525557" exitCode=0 Nov 27 11:57:56 crc kubenswrapper[4796]: I1127 11:57:56.650806 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerDied","Data":"69eb6ccc1e17b4f283f653793de69ab9620fb5164617ca835b2fda897c525557"} Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.499528 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641303 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-scripts\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641399 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-sg-core-conf-yaml\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641476 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-log-httpd\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641508 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-run-httpd\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641560 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjjwt\" (UniqueName: \"kubernetes.io/projected/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-kube-api-access-hjjwt\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641684 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-combined-ca-bundle\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.641701 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-config-data\") pod \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\" (UID: \"ed834c3e-1cc2-488f-a63e-d58ceb3923f6\") " Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.643579 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.647674 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-scripts" (OuterVolumeSpecName: "scripts") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.655356 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.659874 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-kube-api-access-hjjwt" (OuterVolumeSpecName: "kube-api-access-hjjwt") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "kube-api-access-hjjwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.670455 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed834c3e-1cc2-488f-a63e-d58ceb3923f6","Type":"ContainerDied","Data":"e42563294f5e7b145827fb06ec64a10ea9c1001e34612681b0191ea17415d561"} Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.670915 4796 scope.go:117] "RemoveContainer" containerID="0f2e3c2b77e906cfec7058b4a27e9ee75587f66a0a0552f5af67cf5c9a0460e5" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.671578 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.680394 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f94c75f96-tmxbc" event={"ID":"38cfea73-e382-40ee-a600-7f0c340b9e89","Type":"ContainerStarted","Data":"cc48832d297fc306cd59b0e71e47da85ea004205909f4f84af980a1991580dc5"} Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.681651 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.689423 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.710752 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" event={"ID":"74b266b4-8a6e-4b8f-91c8-de16da0f76e8","Type":"ContainerStarted","Data":"4f0a0ba4fb50ef31c83024c5bde1aaa8a3909ce8bbc471cecb40da5ff356c8f7"} Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.711443 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.711614 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-f94c75f96-tmxbc" podStartSLOduration=2.178132473 podStartE2EDuration="4.711591088s" podCreationTimestamp="2025-11-27 11:57:53 +0000 UTC" firstStartedPulling="2025-11-27 11:57:54.321400998 +0000 UTC m=+1991.839719916" lastFinishedPulling="2025-11-27 11:57:56.854859603 +0000 UTC m=+1994.373178531" observedRunningTime="2025-11-27 11:57:57.699425334 +0000 UTC m=+1995.217744252" watchObservedRunningTime="2025-11-27 11:57:57.711591088 +0000 UTC m=+1995.229910006" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.720456 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" event={"ID":"8cdbd3cc-dbee-4abb-8e34-c910554b5142","Type":"ContainerStarted","Data":"049434f9a8690be59e451762d02797a69d0a8251bd0e920c60965444600a5c0b"} Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.721298 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.737403 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" podStartSLOduration=2.5385865 podStartE2EDuration="4.737387176s" podCreationTimestamp="2025-11-27 11:57:53 +0000 UTC" firstStartedPulling="2025-11-27 11:57:54.842068309 +0000 UTC m=+1992.360387227" lastFinishedPulling="2025-11-27 11:57:57.040868985 +0000 UTC m=+1994.559187903" observedRunningTime="2025-11-27 11:57:57.733998356 +0000 UTC m=+1995.252317284" watchObservedRunningTime="2025-11-27 11:57:57.737387176 +0000 UTC m=+1995.255706094" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.743468 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.743496 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.743506 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.743514 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.743525 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjjwt\" (UniqueName: \"kubernetes.io/projected/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-kube-api-access-hjjwt\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.760349 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" podStartSLOduration=4.760336068 podStartE2EDuration="4.760336068s" podCreationTimestamp="2025-11-27 11:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:57:57.759390934 +0000 UTC m=+1995.277709852" watchObservedRunningTime="2025-11-27 11:57:57.760336068 +0000 UTC m=+1995.278654986" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.824086 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-config-data" (OuterVolumeSpecName: "config-data") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.825320 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed834c3e-1cc2-488f-a63e-d58ceb3923f6" (UID: "ed834c3e-1cc2-488f-a63e-d58ceb3923f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.846632 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.846662 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed834c3e-1cc2-488f-a63e-d58ceb3923f6-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.858685 4796 scope.go:117] "RemoveContainer" containerID="967961baae2e08bacffd81376bdbc6f6a9447cddc3481c1f4f37c47db835d6ae" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.885493 4796 scope.go:117] "RemoveContainer" containerID="88121c6cae30fd137788d814fb2e15e2d298a0bd120cc42249dabfaf03062d0e" Nov 27 11:57:57 crc kubenswrapper[4796]: I1127 11:57:57.906957 4796 scope.go:117] "RemoveContainer" containerID="69eb6ccc1e17b4f283f653793de69ab9620fb5164617ca835b2fda897c525557" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.010762 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.019443 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.031829 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:58 crc kubenswrapper[4796]: E1127 11:57:58.032724 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="proxy-httpd" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.032887 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="proxy-httpd" Nov 27 11:57:58 crc kubenswrapper[4796]: E1127 11:57:58.033210 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-notification-agent" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.033396 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-notification-agent" Nov 27 11:57:58 crc kubenswrapper[4796]: E1127 11:57:58.033504 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="sg-core" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.033996 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="sg-core" Nov 27 11:57:58 crc kubenswrapper[4796]: E1127 11:57:58.034094 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-central-agent" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.034150 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-central-agent" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.034472 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-notification-agent" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.034553 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="ceilometer-central-agent" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.034645 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="proxy-httpd" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.034735 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" containerName="sg-core" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.036736 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.039400 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.040052 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.050907 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.153497 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.153785 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-config-data\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.154162 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-log-httpd\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.154326 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-scripts\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.154452 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-run-httpd\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.154563 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lgh8\" (UniqueName: \"kubernetes.io/projected/452c5cce-4b40-4f0b-8975-4c395f8a703f-kube-api-access-4lgh8\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.154707 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.256735 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-config-data\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.256796 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-log-httpd\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.256826 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-scripts\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.256856 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-run-httpd\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.256891 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lgh8\" (UniqueName: \"kubernetes.io/projected/452c5cce-4b40-4f0b-8975-4c395f8a703f-kube-api-access-4lgh8\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.256964 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.257020 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.257406 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-log-httpd\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.257560 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-run-httpd\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.261243 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-scripts\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.263826 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.264010 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.264319 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-config-data\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.276085 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lgh8\" (UniqueName: \"kubernetes.io/projected/452c5cce-4b40-4f0b-8975-4c395f8a703f-kube-api-access-4lgh8\") pod \"ceilometer-0\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.371452 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.903146 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.903801 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-log" containerID="cri-o://378dff5c2638cd67b1de50ad9f7aad86648b95a18567b8218e9ed545b4a26020" gracePeriod=30 Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.903922 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-httpd" containerID="cri-o://df3a543a9ac73d72ec17cbd51fe1220884beb423d558f79abd2ce1abf5c033e5" gracePeriod=30 Nov 27 11:57:58 crc kubenswrapper[4796]: I1127 11:57:58.924884 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.196818 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5469764df4-8wt46"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.200070 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.210125 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5469764df4-8wt46"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.233082 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-f48d67bf7-xvkst"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.234705 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.250191 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6fc467f47d-q8ztt"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.252513 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.257462 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f48d67bf7-xvkst"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.287694 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6fc467f47d-q8ztt"] Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whspp\" (UniqueName: \"kubernetes.io/projected/e3b2e688-5f6c-4ead-9296-88a305dca376-kube-api-access-whspp\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382284 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-combined-ca-bundle\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382325 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382371 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-combined-ca-bundle\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382437 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n49th\" (UniqueName: \"kubernetes.io/projected/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-kube-api-access-n49th\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382488 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-config-data-custom\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382548 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsbwg\" (UniqueName: \"kubernetes.io/projected/ecd41927-ca66-42f8-871b-2b872ba9c3c8-kube-api-access-qsbwg\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382624 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data-custom\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382680 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data-custom\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382708 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-combined-ca-bundle\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382728 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.382761 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-config-data\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485058 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-config-data\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485122 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whspp\" (UniqueName: \"kubernetes.io/projected/e3b2e688-5f6c-4ead-9296-88a305dca376-kube-api-access-whspp\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485147 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-combined-ca-bundle\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485174 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485204 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-combined-ca-bundle\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485237 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n49th\" (UniqueName: \"kubernetes.io/projected/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-kube-api-access-n49th\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485283 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-config-data-custom\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485320 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsbwg\" (UniqueName: \"kubernetes.io/projected/ecd41927-ca66-42f8-871b-2b872ba9c3c8-kube-api-access-qsbwg\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485347 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data-custom\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485383 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data-custom\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485415 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-combined-ca-bundle\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.485431 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.491926 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data-custom\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.492094 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.492353 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-config-data-custom\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.493721 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data-custom\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.493829 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-combined-ca-bundle\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.494857 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-combined-ca-bundle\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.495741 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-combined-ca-bundle\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.499588 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecd41927-ca66-42f8-871b-2b872ba9c3c8-config-data\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.501778 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whspp\" (UniqueName: \"kubernetes.io/projected/e3b2e688-5f6c-4ead-9296-88a305dca376-kube-api-access-whspp\") pod \"heat-cfnapi-f48d67bf7-xvkst\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.506176 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsbwg\" (UniqueName: \"kubernetes.io/projected/ecd41927-ca66-42f8-871b-2b872ba9c3c8-kube-api-access-qsbwg\") pod \"heat-engine-5469764df4-8wt46\" (UID: \"ecd41927-ca66-42f8-871b-2b872ba9c3c8\") " pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.506261 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n49th\" (UniqueName: \"kubernetes.io/projected/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-kube-api-access-n49th\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.506579 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data\") pod \"heat-api-6fc467f47d-q8ztt\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.535334 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.563800 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.577769 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.591471 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed834c3e-1cc2-488f-a63e-d58ceb3923f6" path="/var/lib/kubelet/pods/ed834c3e-1cc2-488f-a63e-d58ceb3923f6/volumes" Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.757026 4796 generic.go:334] "Generic (PLEG): container finished" podID="2960ce96-6701-446e-a800-6f5f9333a322" containerID="378dff5c2638cd67b1de50ad9f7aad86648b95a18567b8218e9ed545b4a26020" exitCode=143 Nov 27 11:57:59 crc kubenswrapper[4796]: I1127 11:57:59.757518 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2960ce96-6701-446e-a800-6f5f9333a322","Type":"ContainerDied","Data":"378dff5c2638cd67b1de50ad9f7aad86648b95a18567b8218e9ed545b4a26020"} Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.580061 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-f94c75f96-tmxbc"] Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.598342 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5d6d658794-wxmmq"] Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.598573 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" podUID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" containerName="heat-cfnapi" containerID="cri-o://4f0a0ba4fb50ef31c83024c5bde1aaa8a3909ce8bbc471cecb40da5ff356c8f7" gracePeriod=60 Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.610587 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6896844f76-vnbq9"] Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.612194 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.619115 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.619423 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.640825 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6896844f76-vnbq9"] Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.675869 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5b7ccbdf6f-g754q"] Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.677249 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.679697 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.680420 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.701415 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b7ccbdf6f-g754q"] Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.713577 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-internal-tls-certs\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.717549 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-config-data\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.717644 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-config-data-custom\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.717744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nrqk\" (UniqueName: \"kubernetes.io/projected/acf45387-6183-41dd-848c-160078149bfd-kube-api-access-7nrqk\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.717819 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-public-tls-certs\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.717853 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-combined-ca-bundle\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.778401 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-f94c75f96-tmxbc" podUID="38cfea73-e382-40ee-a600-7f0c340b9e89" containerName="heat-api" containerID="cri-o://cc48832d297fc306cd59b0e71e47da85ea004205909f4f84af980a1991580dc5" gracePeriod=60 Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819178 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dspq7\" (UniqueName: \"kubernetes.io/projected/6e828f09-14e9-444f-9ef6-6a07f5f665e5-kube-api-access-dspq7\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819246 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-public-tls-certs\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819293 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-config-data-custom\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819334 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-internal-tls-certs\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819403 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-combined-ca-bundle\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819455 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-config-data\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819482 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-config-data\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819519 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-config-data-custom\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819569 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nrqk\" (UniqueName: \"kubernetes.io/projected/acf45387-6183-41dd-848c-160078149bfd-kube-api-access-7nrqk\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819626 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-public-tls-certs\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819656 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-combined-ca-bundle\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.819706 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-internal-tls-certs\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.827484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-combined-ca-bundle\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.827521 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-public-tls-certs\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.827754 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-internal-tls-certs\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.828194 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-config-data\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.828437 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/acf45387-6183-41dd-848c-160078149bfd-config-data-custom\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.846354 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nrqk\" (UniqueName: \"kubernetes.io/projected/acf45387-6183-41dd-848c-160078149bfd-kube-api-access-7nrqk\") pod \"heat-api-6896844f76-vnbq9\" (UID: \"acf45387-6183-41dd-848c-160078149bfd\") " pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.921748 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-internal-tls-certs\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.921800 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dspq7\" (UniqueName: \"kubernetes.io/projected/6e828f09-14e9-444f-9ef6-6a07f5f665e5-kube-api-access-dspq7\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.921826 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-public-tls-certs\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.921846 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-config-data-custom\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.921898 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-combined-ca-bundle\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.921934 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-config-data\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.926385 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-config-data\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.926416 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-config-data-custom\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.927037 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-public-tls-certs\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.927743 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-internal-tls-certs\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.939496 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.941020 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e828f09-14e9-444f-9ef6-6a07f5f665e5-combined-ca-bundle\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.945933 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dspq7\" (UniqueName: \"kubernetes.io/projected/6e828f09-14e9-444f-9ef6-6a07f5f665e5-kube-api-access-dspq7\") pod \"heat-cfnapi-5b7ccbdf6f-g754q\" (UID: \"6e828f09-14e9-444f-9ef6-6a07f5f665e5\") " pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:00 crc kubenswrapper[4796]: I1127 11:58:00.999367 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:01 crc kubenswrapper[4796]: I1127 11:58:01.563562 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:01 crc kubenswrapper[4796]: I1127 11:58:01.788887 4796 generic.go:334] "Generic (PLEG): container finished" podID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" containerID="4f0a0ba4fb50ef31c83024c5bde1aaa8a3909ce8bbc471cecb40da5ff356c8f7" exitCode=0 Nov 27 11:58:01 crc kubenswrapper[4796]: I1127 11:58:01.788956 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" event={"ID":"74b266b4-8a6e-4b8f-91c8-de16da0f76e8","Type":"ContainerDied","Data":"4f0a0ba4fb50ef31c83024c5bde1aaa8a3909ce8bbc471cecb40da5ff356c8f7"} Nov 27 11:58:01 crc kubenswrapper[4796]: I1127 11:58:01.791118 4796 generic.go:334] "Generic (PLEG): container finished" podID="38cfea73-e382-40ee-a600-7f0c340b9e89" containerID="cc48832d297fc306cd59b0e71e47da85ea004205909f4f84af980a1991580dc5" exitCode=0 Nov 27 11:58:01 crc kubenswrapper[4796]: I1127 11:58:01.791162 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f94c75f96-tmxbc" event={"ID":"38cfea73-e382-40ee-a600-7f0c340b9e89","Type":"ContainerDied","Data":"cc48832d297fc306cd59b0e71e47da85ea004205909f4f84af980a1991580dc5"} Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.420253 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.420766 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-log" containerID="cri-o://8acd37e75f4453b24d2135a0ebba5b36b522d63e4fc64354e61a2026c4c1d889" gracePeriod=30 Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.420943 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-httpd" containerID="cri-o://e8fb4261492dd9c8fb3150ffa8de4241e497c1e9615dfe5b802ed25978cdb786" gracePeriod=30 Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.807814 4796 generic.go:334] "Generic (PLEG): container finished" podID="2960ce96-6701-446e-a800-6f5f9333a322" containerID="df3a543a9ac73d72ec17cbd51fe1220884beb423d558f79abd2ce1abf5c033e5" exitCode=0 Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.808690 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2960ce96-6701-446e-a800-6f5f9333a322","Type":"ContainerDied","Data":"df3a543a9ac73d72ec17cbd51fe1220884beb423d558f79abd2ce1abf5c033e5"} Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.814100 4796 generic.go:334] "Generic (PLEG): container finished" podID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerID="8acd37e75f4453b24d2135a0ebba5b36b522d63e4fc64354e61a2026c4c1d889" exitCode=143 Nov 27 11:58:02 crc kubenswrapper[4796]: I1127 11:58:02.814162 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"453af780-fb5a-4048-b9fd-2ce7a497916c","Type":"ContainerDied","Data":"8acd37e75f4453b24d2135a0ebba5b36b522d63e4fc64354e61a2026c4c1d889"} Nov 27 11:58:03 crc kubenswrapper[4796]: I1127 11:58:03.770368 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:58:03 crc kubenswrapper[4796]: I1127 11:58:03.782906 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-f94c75f96-tmxbc" podUID="38cfea73-e382-40ee-a600-7f0c340b9e89" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.180:8004/healthcheck\": dial tcp 10.217.0.180:8004: connect: connection refused" Nov 27 11:58:03 crc kubenswrapper[4796]: I1127 11:58:03.838567 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-27ssb"] Nov 27 11:58:03 crc kubenswrapper[4796]: I1127 11:58:03.838797 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerName="dnsmasq-dns" containerID="cri-o://b4b7555a272a493026f39d1be9a5119c131b08da4f9989e0317c66e9eb6a6850" gracePeriod=10 Nov 27 11:58:04 crc kubenswrapper[4796]: I1127 11:58:04.160215 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" podUID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.181:8000/healthcheck\": dial tcp 10.217.0.181:8000: connect: connection refused" Nov 27 11:58:04 crc kubenswrapper[4796]: I1127 11:58:04.852398 4796 generic.go:334] "Generic (PLEG): container finished" podID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerID="b4b7555a272a493026f39d1be9a5119c131b08da4f9989e0317c66e9eb6a6850" exitCode=0 Nov 27 11:58:04 crc kubenswrapper[4796]: I1127 11:58:04.852657 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" event={"ID":"4f58631f-4476-46a4-969f-e6d7c989fb3e","Type":"ContainerDied","Data":"b4b7555a272a493026f39d1be9a5119c131b08da4f9989e0317c66e9eb6a6850"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.066203 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.109750 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.122542 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.127175 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213081 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-config\") pod \"4f58631f-4476-46a4-969f-e6d7c989fb3e\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213148 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4wjk\" (UniqueName: \"kubernetes.io/projected/4f58631f-4476-46a4-969f-e6d7c989fb3e-kube-api-access-p4wjk\") pod \"4f58631f-4476-46a4-969f-e6d7c989fb3e\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213195 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-sb\") pod \"4f58631f-4476-46a4-969f-e6d7c989fb3e\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213237 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-config-data\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213262 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-combined-ca-bundle\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213315 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-combined-ca-bundle\") pod \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213373 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr9xs\" (UniqueName: \"kubernetes.io/projected/2960ce96-6701-446e-a800-6f5f9333a322-kube-api-access-pr9xs\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213419 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-svc\") pod \"4f58631f-4476-46a4-969f-e6d7c989fb3e\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213470 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data\") pod \"38cfea73-e382-40ee-a600-7f0c340b9e89\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213506 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-scripts\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213534 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-combined-ca-bundle\") pod \"38cfea73-e382-40ee-a600-7f0c340b9e89\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213560 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213584 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-nb\") pod \"4f58631f-4476-46a4-969f-e6d7c989fb3e\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213619 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8w4l\" (UniqueName: \"kubernetes.io/projected/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-kube-api-access-h8w4l\") pod \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213642 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data-custom\") pod \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213687 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-logs\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213720 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-httpd-run\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213759 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data\") pod \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\" (UID: \"74b266b4-8a6e-4b8f-91c8-de16da0f76e8\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213788 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzs2r\" (UniqueName: \"kubernetes.io/projected/38cfea73-e382-40ee-a600-7f0c340b9e89-kube-api-access-bzs2r\") pod \"38cfea73-e382-40ee-a600-7f0c340b9e89\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213812 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-swift-storage-0\") pod \"4f58631f-4476-46a4-969f-e6d7c989fb3e\" (UID: \"4f58631f-4476-46a4-969f-e6d7c989fb3e\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213873 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-public-tls-certs\") pod \"2960ce96-6701-446e-a800-6f5f9333a322\" (UID: \"2960ce96-6701-446e-a800-6f5f9333a322\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.213930 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data-custom\") pod \"38cfea73-e382-40ee-a600-7f0c340b9e89\" (UID: \"38cfea73-e382-40ee-a600-7f0c340b9e89\") " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.226009 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.226384 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "38cfea73-e382-40ee-a600-7f0c340b9e89" (UID: "38cfea73-e382-40ee-a600-7f0c340b9e89"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.233563 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.238899 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-logs" (OuterVolumeSpecName: "logs") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.244974 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f58631f-4476-46a4-969f-e6d7c989fb3e-kube-api-access-p4wjk" (OuterVolumeSpecName: "kube-api-access-p4wjk") pod "4f58631f-4476-46a4-969f-e6d7c989fb3e" (UID: "4f58631f-4476-46a4-969f-e6d7c989fb3e"). InnerVolumeSpecName "kube-api-access-p4wjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.248801 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "74b266b4-8a6e-4b8f-91c8-de16da0f76e8" (UID: "74b266b4-8a6e-4b8f-91c8-de16da0f76e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.250430 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-scripts" (OuterVolumeSpecName: "scripts") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.259457 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2960ce96-6701-446e-a800-6f5f9333a322-kube-api-access-pr9xs" (OuterVolumeSpecName: "kube-api-access-pr9xs") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "kube-api-access-pr9xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.259820 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-kube-api-access-h8w4l" (OuterVolumeSpecName: "kube-api-access-h8w4l") pod "74b266b4-8a6e-4b8f-91c8-de16da0f76e8" (UID: "74b266b4-8a6e-4b8f-91c8-de16da0f76e8"). InnerVolumeSpecName "kube-api-access-h8w4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.269903 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38cfea73-e382-40ee-a600-7f0c340b9e89-kube-api-access-bzs2r" (OuterVolumeSpecName: "kube-api-access-bzs2r") pod "38cfea73-e382-40ee-a600-7f0c340b9e89" (UID: "38cfea73-e382-40ee-a600-7f0c340b9e89"). InnerVolumeSpecName "kube-api-access-bzs2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320310 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320340 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4wjk\" (UniqueName: \"kubernetes.io/projected/4f58631f-4476-46a4-969f-e6d7c989fb3e-kube-api-access-p4wjk\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320352 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr9xs\" (UniqueName: \"kubernetes.io/projected/2960ce96-6701-446e-a800-6f5f9333a322-kube-api-access-pr9xs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320421 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320444 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320453 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8w4l\" (UniqueName: \"kubernetes.io/projected/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-kube-api-access-h8w4l\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320462 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320471 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320479 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2960ce96-6701-446e-a800-6f5f9333a322-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.320488 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzs2r\" (UniqueName: \"kubernetes.io/projected/38cfea73-e382-40ee-a600-7f0c340b9e89-kube-api-access-bzs2r\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.425856 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74b266b4-8a6e-4b8f-91c8-de16da0f76e8" (UID: "74b266b4-8a6e-4b8f-91c8-de16da0f76e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.428917 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.435544 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data" (OuterVolumeSpecName: "config-data") pod "74b266b4-8a6e-4b8f-91c8-de16da0f76e8" (UID: "74b266b4-8a6e-4b8f-91c8-de16da0f76e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.444617 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data" (OuterVolumeSpecName: "config-data") pod "38cfea73-e382-40ee-a600-7f0c340b9e89" (UID: "38cfea73-e382-40ee-a600-7f0c340b9e89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.469709 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.485360 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f58631f-4476-46a4-969f-e6d7c989fb3e" (UID: "4f58631f-4476-46a4-969f-e6d7c989fb3e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.493669 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.502327 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-config-data" (OuterVolumeSpecName: "config-data") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.502415 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38cfea73-e382-40ee-a600-7f0c340b9e89" (UID: "38cfea73-e382-40ee-a600-7f0c340b9e89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.502746 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-config" (OuterVolumeSpecName: "config") pod "4f58631f-4476-46a4-969f-e6d7c989fb3e" (UID: "4f58631f-4476-46a4-969f-e6d7c989fb3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.507240 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f58631f-4476-46a4-969f-e6d7c989fb3e" (UID: "4f58631f-4476-46a4-969f-e6d7c989fb3e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.525997 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4f58631f-4476-46a4-969f-e6d7c989fb3e" (UID: "4f58631f-4476-46a4-969f-e6d7c989fb3e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540037 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540460 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540471 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38cfea73-e382-40ee-a600-7f0c340b9e89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540483 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540493 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b266b4-8a6e-4b8f-91c8-de16da0f76e8-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540514 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540523 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540531 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540539 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.540547 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.556168 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f58631f-4476-46a4-969f-e6d7c989fb3e" (UID: "4f58631f-4476-46a4-969f-e6d7c989fb3e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.559436 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2960ce96-6701-446e-a800-6f5f9333a322" (UID: "2960ce96-6701-446e-a800-6f5f9333a322"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.569739 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:58:05 crc kubenswrapper[4796]: E1127 11:58:05.570342 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.643226 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f58631f-4476-46a4-969f-e6d7c989fb3e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.643312 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2960ce96-6701-446e-a800-6f5f9333a322-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.716704 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b7ccbdf6f-g754q"] Nov 27 11:58:05 crc kubenswrapper[4796]: W1127 11:58:05.756266 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e828f09_14e9_444f_9ef6_6a07f5f665e5.slice/crio-a070c477b735b0b5b23eff0037cd55a6cfb2d2b17a5fa1171e8a62c87f24c0eb WatchSource:0}: Error finding container a070c477b735b0b5b23eff0037cd55a6cfb2d2b17a5fa1171e8a62c87f24c0eb: Status 404 returned error can't find the container with id a070c477b735b0b5b23eff0037cd55a6cfb2d2b17a5fa1171e8a62c87f24c0eb Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.881447 4796 generic.go:334] "Generic (PLEG): container finished" podID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerID="e8fb4261492dd9c8fb3150ffa8de4241e497c1e9615dfe5b802ed25978cdb786" exitCode=0 Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.881553 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"453af780-fb5a-4048-b9fd-2ce7a497916c","Type":"ContainerDied","Data":"e8fb4261492dd9c8fb3150ffa8de4241e497c1e9615dfe5b802ed25978cdb786"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.900597 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2960ce96-6701-446e-a800-6f5f9333a322","Type":"ContainerDied","Data":"2b8ad5f98f8adf10585dacc5ef003217186af7733adf74230db95a1ec7b8aa5b"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.900658 4796 scope.go:117] "RemoveContainer" containerID="df3a543a9ac73d72ec17cbd51fe1220884beb423d558f79abd2ce1abf5c033e5" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.900828 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.909452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" event={"ID":"6e828f09-14e9-444f-9ef6-6a07f5f665e5","Type":"ContainerStarted","Data":"a070c477b735b0b5b23eff0037cd55a6cfb2d2b17a5fa1171e8a62c87f24c0eb"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.912203 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" event={"ID":"4f58631f-4476-46a4-969f-e6d7c989fb3e","Type":"ContainerDied","Data":"6ea78b6de89439bf36325d2cd5ea143f6673e2c167d47d1070733062ac7c7048"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.912260 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-27ssb" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.916550 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-whbbq" event={"ID":"6524424e-afc2-443c-beac-c0503ac10ddf","Type":"ContainerStarted","Data":"064006ab2a9695d2a447ea377d0daf5fc0c97676ebc78389fc16ebd4dd27c40d"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.932495 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f94c75f96-tmxbc" event={"ID":"38cfea73-e382-40ee-a600-7f0c340b9e89","Type":"ContainerDied","Data":"7abd9a975dbcacddab6b30e3a5ce60ccccb3783f25852eccc373cf3132a22f59"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.932622 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f94c75f96-tmxbc" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.942303 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.960588 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" event={"ID":"74b266b4-8a6e-4b8f-91c8-de16da0f76e8","Type":"ContainerDied","Data":"530c5c14306c324f15ffa0dde4472e322e16e46bae7cd304c19a0559b19803ab"} Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.960705 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5d6d658794-wxmmq" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.979909 4796 scope.go:117] "RemoveContainer" containerID="378dff5c2638cd67b1de50ad9f7aad86648b95a18567b8218e9ed545b4a26020" Nov 27 11:58:05 crc kubenswrapper[4796]: I1127 11:58:05.997512 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.003223 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerStarted","Data":"593b8a46f9c3ec5377b6dcfa338c3e2fe91a8b9e4e881611b0f04ccfa06097be"} Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.003272 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerStarted","Data":"10ed894bf6adb736717458dccb31bfb5e15a5cc36dbec5c5f329cc14ab0dc3bf"} Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.011669 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-27ssb"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.049235 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-27ssb"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.065666 4796 scope.go:117] "RemoveContainer" containerID="b4b7555a272a493026f39d1be9a5119c131b08da4f9989e0317c66e9eb6a6850" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085070 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:58:06 crc kubenswrapper[4796]: E1127 11:58:06.085499 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38cfea73-e382-40ee-a600-7f0c340b9e89" containerName="heat-api" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085518 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="38cfea73-e382-40ee-a600-7f0c340b9e89" containerName="heat-api" Nov 27 11:58:06 crc kubenswrapper[4796]: E1127 11:58:06.085534 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerName="dnsmasq-dns" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085539 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerName="dnsmasq-dns" Nov 27 11:58:06 crc kubenswrapper[4796]: E1127 11:58:06.085563 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-log" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085569 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-log" Nov 27 11:58:06 crc kubenswrapper[4796]: E1127 11:58:06.085585 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-httpd" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085591 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-httpd" Nov 27 11:58:06 crc kubenswrapper[4796]: E1127 11:58:06.085603 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerName="init" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085609 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerName="init" Nov 27 11:58:06 crc kubenswrapper[4796]: E1127 11:58:06.085623 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" containerName="heat-cfnapi" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085628 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" containerName="heat-cfnapi" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085803 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" containerName="heat-cfnapi" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085820 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="38cfea73-e382-40ee-a600-7f0c340b9e89" containerName="heat-api" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085832 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-log" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085846 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="2960ce96-6701-446e-a800-6f5f9333a322" containerName="glance-httpd" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.085865 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" containerName="dnsmasq-dns" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.086762 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.091593 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.091881 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.097002 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-whbbq" podStartSLOduration=2.263443712 podStartE2EDuration="14.096976514s" podCreationTimestamp="2025-11-27 11:57:52 +0000 UTC" firstStartedPulling="2025-11-27 11:57:53.248823985 +0000 UTC m=+1990.767142903" lastFinishedPulling="2025-11-27 11:58:05.082356777 +0000 UTC m=+2002.600675705" observedRunningTime="2025-11-27 11:58:05.995623221 +0000 UTC m=+2003.513942139" watchObservedRunningTime="2025-11-27 11:58:06.096976514 +0000 UTC m=+2003.615295432" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.099409 4796 scope.go:117] "RemoveContainer" containerID="b89f7f5e738b467e6b0a50a7f6a595d074233ecedb972e35279d6ebf2b5eb6ea" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.129955 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.141402 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5d6d658794-wxmmq"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.154409 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5d6d658794-wxmmq"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.165514 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-f94c75f96-tmxbc"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.173083 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-f94c75f96-tmxbc"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.182332 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f48d67bf7-xvkst"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.197421 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6fc467f47d-q8ztt"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.198944 4796 scope.go:117] "RemoveContainer" containerID="cc48832d297fc306cd59b0e71e47da85ea004205909f4f84af980a1991580dc5" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.199520 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6896844f76-vnbq9"] Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.210043 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5469764df4-8wt46"] Nov 27 11:58:06 crc kubenswrapper[4796]: W1127 11:58:06.230651 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fd6de03_a43e_44b2_97d0_ac73cf877eeb.slice/crio-d729e7755481800c9fc46f6cba308b6c12752390a7d490b22fb37af9079474a8 WatchSource:0}: Error finding container d729e7755481800c9fc46f6cba308b6c12752390a7d490b22fb37af9079474a8: Status 404 returned error can't find the container with id d729e7755481800c9fc46f6cba308b6c12752390a7d490b22fb37af9079474a8 Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.256655 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.258299 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.258803 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.259407 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.259653 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zqp5\" (UniqueName: \"kubernetes.io/projected/1f85834f-7915-4c31-b55d-5c4948763a3c-kube-api-access-2zqp5\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.259893 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f85834f-7915-4c31-b55d-5c4948763a3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.262007 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f85834f-7915-4c31-b55d-5c4948763a3c-logs\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.262294 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.363738 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364044 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364073 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364104 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zqp5\" (UniqueName: \"kubernetes.io/projected/1f85834f-7915-4c31-b55d-5c4948763a3c-kube-api-access-2zqp5\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364149 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f85834f-7915-4c31-b55d-5c4948763a3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364199 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f85834f-7915-4c31-b55d-5c4948763a3c-logs\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364244 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.364294 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.365712 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.372079 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f85834f-7915-4c31-b55d-5c4948763a3c-logs\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.372399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f85834f-7915-4c31-b55d-5c4948763a3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.378252 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.382240 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.383164 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.384075 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f85834f-7915-4c31-b55d-5c4948763a3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.409447 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zqp5\" (UniqueName: \"kubernetes.io/projected/1f85834f-7915-4c31-b55d-5c4948763a3c-kube-api-access-2zqp5\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.420474 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1f85834f-7915-4c31-b55d-5c4948763a3c\") " pod="openstack/glance-default-external-api-0" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.466637 4796 scope.go:117] "RemoveContainer" containerID="4f0a0ba4fb50ef31c83024c5bde1aaa8a3909ce8bbc471cecb40da5ff356c8f7" Nov 27 11:58:06 crc kubenswrapper[4796]: I1127 11:58:06.713734 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.049387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc467f47d-q8ztt" event={"ID":"8fd6de03-a43e-44b2-97d0-ac73cf877eeb","Type":"ContainerStarted","Data":"2d2ea4beed51f74ec71177d9a8271bc5bc6bd8dab10c71f9423d2cba9384bf2d"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.049736 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc467f47d-q8ztt" event={"ID":"8fd6de03-a43e-44b2-97d0-ac73cf877eeb","Type":"ContainerStarted","Data":"d729e7755481800c9fc46f6cba308b6c12752390a7d490b22fb37af9079474a8"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.049954 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.072755 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6896844f76-vnbq9" event={"ID":"acf45387-6183-41dd-848c-160078149bfd","Type":"ContainerStarted","Data":"b3684e5359d09398c66666b540da8d3903a72b2e50a87bedc6792b9b90d9ac29"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.072807 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6896844f76-vnbq9" event={"ID":"acf45387-6183-41dd-848c-160078149bfd","Type":"ContainerStarted","Data":"bb836d814dd13907cf8bbf857fb3a43124dce78ea414864563c8bd56103ac5bf"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.073674 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.099786 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" event={"ID":"e3b2e688-5f6c-4ead-9296-88a305dca376","Type":"ContainerStarted","Data":"b4a72c7c66094d81c56f059d250945675c74dc7fc3148fb61e9f99c31fadb4f9"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.100022 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" event={"ID":"e3b2e688-5f6c-4ead-9296-88a305dca376","Type":"ContainerStarted","Data":"1dba3b6bfda18e171d37e8415d52456542a5e0982fc856bf91e689bd0caa23ee"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.103627 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.130945 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6896844f76-vnbq9" podStartSLOduration=7.130923876 podStartE2EDuration="7.130923876s" podCreationTimestamp="2025-11-27 11:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:07.10182046 +0000 UTC m=+2004.620139378" watchObservedRunningTime="2025-11-27 11:58:07.130923876 +0000 UTC m=+2004.649242794" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.139073 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6fc467f47d-q8ztt" podStartSLOduration=8.139053824 podStartE2EDuration="8.139053824s" podCreationTimestamp="2025-11-27 11:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:07.083007418 +0000 UTC m=+2004.601326336" watchObservedRunningTime="2025-11-27 11:58:07.139053824 +0000 UTC m=+2004.657372742" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.140164 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5469764df4-8wt46" event={"ID":"ecd41927-ca66-42f8-871b-2b872ba9c3c8","Type":"ContainerStarted","Data":"e06b65ebec01a3109306fb7cf29d6da06270711f958efb84c13e168a83be93b9"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.140219 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5469764df4-8wt46" event={"ID":"ecd41927-ca66-42f8-871b-2b872ba9c3c8","Type":"ContainerStarted","Data":"2da9c9f4a0ef08e91ea1225f2246f957bf5b9cbec327d78e378db31d04a23f2e"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.140294 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.143067 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" podStartSLOduration=8.143049601 podStartE2EDuration="8.143049601s" podCreationTimestamp="2025-11-27 11:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:07.121117775 +0000 UTC m=+2004.639436693" watchObservedRunningTime="2025-11-27 11:58:07.143049601 +0000 UTC m=+2004.661368519" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.143952 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" event={"ID":"6e828f09-14e9-444f-9ef6-6a07f5f665e5","Type":"ContainerStarted","Data":"769f1bfed54d3cc6235087e4bbd9314cbe5e9df752b6b9cb52f42ffcd64d7a83"} Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.143981 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.165172 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5469764df4-8wt46" podStartSLOduration=8.16515328 podStartE2EDuration="8.16515328s" podCreationTimestamp="2025-11-27 11:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:07.164765529 +0000 UTC m=+2004.683084457" watchObservedRunningTime="2025-11-27 11:58:07.16515328 +0000 UTC m=+2004.683472188" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.212378 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" podStartSLOduration=7.212328128 podStartE2EDuration="7.212328128s" podCreationTimestamp="2025-11-27 11:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:07.201435557 +0000 UTC m=+2004.719754485" watchObservedRunningTime="2025-11-27 11:58:07.212328128 +0000 UTC m=+2004.730647056" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.416942 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.518930 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.584976 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2960ce96-6701-446e-a800-6f5f9333a322" path="/var/lib/kubelet/pods/2960ce96-6701-446e-a800-6f5f9333a322/volumes" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.590249 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38cfea73-e382-40ee-a600-7f0c340b9e89" path="/var/lib/kubelet/pods/38cfea73-e382-40ee-a600-7f0c340b9e89/volumes" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.591097 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f58631f-4476-46a4-969f-e6d7c989fb3e" path="/var/lib/kubelet/pods/4f58631f-4476-46a4-969f-e6d7c989fb3e/volumes" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.591778 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b266b4-8a6e-4b8f-91c8-de16da0f76e8" path="/var/lib/kubelet/pods/74b266b4-8a6e-4b8f-91c8-de16da0f76e8/volumes" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614217 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-logs\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614342 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fshkp\" (UniqueName: \"kubernetes.io/projected/453af780-fb5a-4048-b9fd-2ce7a497916c-kube-api-access-fshkp\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614383 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-internal-tls-certs\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614422 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-config-data\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614479 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614500 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-combined-ca-bundle\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614549 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-scripts\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.614564 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-httpd-run\") pod \"453af780-fb5a-4048-b9fd-2ce7a497916c\" (UID: \"453af780-fb5a-4048-b9fd-2ce7a497916c\") " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.615455 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.615703 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-logs" (OuterVolumeSpecName: "logs") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.624493 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/453af780-fb5a-4048-b9fd-2ce7a497916c-kube-api-access-fshkp" (OuterVolumeSpecName: "kube-api-access-fshkp") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "kube-api-access-fshkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.624835 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.639862 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-scripts" (OuterVolumeSpecName: "scripts") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.667241 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.691183 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-config-data" (OuterVolumeSpecName: "config-data") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.716576 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.717284 4796 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.717400 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/453af780-fb5a-4048-b9fd-2ce7a497916c-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.717464 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fshkp\" (UniqueName: \"kubernetes.io/projected/453af780-fb5a-4048-b9fd-2ce7a497916c-kube-api-access-fshkp\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.717524 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.717729 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.717767 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.743891 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.759089 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "453af780-fb5a-4048-b9fd-2ce7a497916c" (UID: "453af780-fb5a-4048-b9fd-2ce7a497916c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.819434 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/453af780-fb5a-4048-b9fd-2ce7a497916c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:07 crc kubenswrapper[4796]: I1127 11:58:07.819476 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.155638 4796 generic.go:334] "Generic (PLEG): container finished" podID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerID="2d2ea4beed51f74ec71177d9a8271bc5bc6bd8dab10c71f9423d2cba9384bf2d" exitCode=1 Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.156011 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc467f47d-q8ztt" event={"ID":"8fd6de03-a43e-44b2-97d0-ac73cf877eeb","Type":"ContainerDied","Data":"2d2ea4beed51f74ec71177d9a8271bc5bc6bd8dab10c71f9423d2cba9384bf2d"} Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.156708 4796 scope.go:117] "RemoveContainer" containerID="2d2ea4beed51f74ec71177d9a8271bc5bc6bd8dab10c71f9423d2cba9384bf2d" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.160232 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1f85834f-7915-4c31-b55d-5c4948763a3c","Type":"ContainerStarted","Data":"d2e62aec6a4f2011ab05aa85212163f7ca60d31e1b639d65605dfd2a7f1662ef"} Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.172479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerStarted","Data":"96b61b12f6b0b5ac9895c143d42c70c802b66089984ee792461cac0c5a45d221"} Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.172525 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerStarted","Data":"4b596e5b2fa264d353b58566df2213d77eb6c2b768939ed737dfbbdf3fa43c8a"} Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.175634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"453af780-fb5a-4048-b9fd-2ce7a497916c","Type":"ContainerDied","Data":"480c745046ed9b4766f57c3c718ea1cedf1e084dff42b2e41053fc17d0ce460c"} Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.175880 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.175959 4796 scope.go:117] "RemoveContainer" containerID="e8fb4261492dd9c8fb3150ffa8de4241e497c1e9615dfe5b802ed25978cdb786" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.177392 4796 generic.go:334] "Generic (PLEG): container finished" podID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerID="b4a72c7c66094d81c56f059d250945675c74dc7fc3148fb61e9f99c31fadb4f9" exitCode=1 Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.178161 4796 scope.go:117] "RemoveContainer" containerID="b4a72c7c66094d81c56f059d250945675c74dc7fc3148fb61e9f99c31fadb4f9" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.178573 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" event={"ID":"e3b2e688-5f6c-4ead-9296-88a305dca376","Type":"ContainerDied","Data":"b4a72c7c66094d81c56f059d250945675c74dc7fc3148fb61e9f99c31fadb4f9"} Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.268616 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.294350 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.302356 4796 scope.go:117] "RemoveContainer" containerID="8acd37e75f4453b24d2135a0ebba5b36b522d63e4fc64354e61a2026c4c1d889" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.319082 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:58:08 crc kubenswrapper[4796]: E1127 11:58:08.319558 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-httpd" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.319572 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-httpd" Nov 27 11:58:08 crc kubenswrapper[4796]: E1127 11:58:08.319599 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-log" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.319605 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-log" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.322684 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-httpd" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.322716 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" containerName="glance-log" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.323711 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.337186 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.339891 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.340727 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.457792 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7681c6a-f516-4b70-848c-2a6ce2c2e087-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.457869 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.457890 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7681c6a-f516-4b70-848c-2a6ce2c2e087-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.457906 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.457937 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.457955 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.458016 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ldtl\" (UniqueName: \"kubernetes.io/projected/f7681c6a-f516-4b70-848c-2a6ce2c2e087-kube-api-access-9ldtl\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.458044 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.559524 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ldtl\" (UniqueName: \"kubernetes.io/projected/f7681c6a-f516-4b70-848c-2a6ce2c2e087-kube-api-access-9ldtl\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.559856 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.559894 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7681c6a-f516-4b70-848c-2a6ce2c2e087-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.559967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.559993 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7681c6a-f516-4b70-848c-2a6ce2c2e087-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.560012 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.560047 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.560068 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.560491 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.562222 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7681c6a-f516-4b70-848c-2a6ce2c2e087-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.563166 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7681c6a-f516-4b70-848c-2a6ce2c2e087-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.572511 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.574173 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.575354 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.575538 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7681c6a-f516-4b70-848c-2a6ce2c2e087-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.583712 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ldtl\" (UniqueName: \"kubernetes.io/projected/f7681c6a-f516-4b70-848c-2a6ce2c2e087-kube-api-access-9ldtl\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.609918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7681c6a-f516-4b70-848c-2a6ce2c2e087\") " pod="openstack/glance-default-internal-api-0" Nov 27 11:58:08 crc kubenswrapper[4796]: I1127 11:58:08.844699 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.233876 4796 generic.go:334] "Generic (PLEG): container finished" podID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerID="6691b124c02b21d7d108860d721432fbf98c9815031c7b2ebb1a7f93017d0cb0" exitCode=1 Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.234656 4796 scope.go:117] "RemoveContainer" containerID="6691b124c02b21d7d108860d721432fbf98c9815031c7b2ebb1a7f93017d0cb0" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.234679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" event={"ID":"e3b2e688-5f6c-4ead-9296-88a305dca376","Type":"ContainerDied","Data":"6691b124c02b21d7d108860d721432fbf98c9815031c7b2ebb1a7f93017d0cb0"} Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.234718 4796 scope.go:117] "RemoveContainer" containerID="b4a72c7c66094d81c56f059d250945675c74dc7fc3148fb61e9f99c31fadb4f9" Nov 27 11:58:09 crc kubenswrapper[4796]: E1127 11:58:09.235053 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f48d67bf7-xvkst_openstack(e3b2e688-5f6c-4ead-9296-88a305dca376)\"" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.247801 4796 generic.go:334] "Generic (PLEG): container finished" podID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerID="f98b7a273c0acc71226485dbf4ac219d4ee6179d3ef20e09b2e8977c69c4cc80" exitCode=1 Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.247886 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc467f47d-q8ztt" event={"ID":"8fd6de03-a43e-44b2-97d0-ac73cf877eeb","Type":"ContainerDied","Data":"f98b7a273c0acc71226485dbf4ac219d4ee6179d3ef20e09b2e8977c69c4cc80"} Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.248512 4796 scope.go:117] "RemoveContainer" containerID="f98b7a273c0acc71226485dbf4ac219d4ee6179d3ef20e09b2e8977c69c4cc80" Nov 27 11:58:09 crc kubenswrapper[4796]: E1127 11:58:09.248730 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6fc467f47d-q8ztt_openstack(8fd6de03-a43e-44b2-97d0-ac73cf877eeb)\"" pod="openstack/heat-api-6fc467f47d-q8ztt" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.268701 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1f85834f-7915-4c31-b55d-5c4948763a3c","Type":"ContainerStarted","Data":"9605113d0becd7267ab19b14580de8a0b3f411b052509c4456a1bccbdf2bdeec"} Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.268759 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1f85834f-7915-4c31-b55d-5c4948763a3c","Type":"ContainerStarted","Data":"dc8ad2ac3cd894d1cc21ded3c1b6d6846d597ec2facb63cf94da3e671ee877db"} Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.361598 4796 scope.go:117] "RemoveContainer" containerID="2d2ea4beed51f74ec71177d9a8271bc5bc6bd8dab10c71f9423d2cba9384bf2d" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.366038 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.366020632 podStartE2EDuration="4.366020632s" podCreationTimestamp="2025-11-27 11:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:09.303353451 +0000 UTC m=+2006.821672369" watchObservedRunningTime="2025-11-27 11:58:09.366020632 +0000 UTC m=+2006.884339550" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.431356 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.563963 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.564303 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.590554 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="453af780-fb5a-4048-b9fd-2ce7a497916c" path="/var/lib/kubelet/pods/453af780-fb5a-4048-b9fd-2ce7a497916c/volumes" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.592103 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:58:09 crc kubenswrapper[4796]: I1127 11:58:09.592229 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.281464 4796 scope.go:117] "RemoveContainer" containerID="f98b7a273c0acc71226485dbf4ac219d4ee6179d3ef20e09b2e8977c69c4cc80" Nov 27 11:58:10 crc kubenswrapper[4796]: E1127 11:58:10.282630 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6fc467f47d-q8ztt_openstack(8fd6de03-a43e-44b2-97d0-ac73cf877eeb)\"" pod="openstack/heat-api-6fc467f47d-q8ztt" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.287629 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerStarted","Data":"cea66ec29760c6fcb188ebf6db18b5a1ed51b73cd112a5af78186041a7ed3f50"} Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.287840 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-central-agent" containerID="cri-o://593b8a46f9c3ec5377b6dcfa338c3e2fe91a8b9e4e881611b0f04ccfa06097be" gracePeriod=30 Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.288299 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.288447 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="sg-core" containerID="cri-o://96b61b12f6b0b5ac9895c143d42c70c802b66089984ee792461cac0c5a45d221" gracePeriod=30 Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.288561 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-notification-agent" containerID="cri-o://4b596e5b2fa264d353b58566df2213d77eb6c2b768939ed737dfbbdf3fa43c8a" gracePeriod=30 Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.288826 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="proxy-httpd" containerID="cri-o://cea66ec29760c6fcb188ebf6db18b5a1ed51b73cd112a5af78186041a7ed3f50" gracePeriod=30 Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.297607 4796 scope.go:117] "RemoveContainer" containerID="6691b124c02b21d7d108860d721432fbf98c9815031c7b2ebb1a7f93017d0cb0" Nov 27 11:58:10 crc kubenswrapper[4796]: E1127 11:58:10.297815 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f48d67bf7-xvkst_openstack(e3b2e688-5f6c-4ead-9296-88a305dca376)\"" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.310970 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7681c6a-f516-4b70-848c-2a6ce2c2e087","Type":"ContainerStarted","Data":"3ccd5de36f5f0f9ea9086d4c787af90f0d158db242a31d556b91ef035e87267e"} Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.311073 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7681c6a-f516-4b70-848c-2a6ce2c2e087","Type":"ContainerStarted","Data":"f6af0fa27247846390ad490fb3ab29a54f23d80d0a1cf8a2c619313c0ff35971"} Nov 27 11:58:10 crc kubenswrapper[4796]: I1127 11:58:10.340869 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.069129951 podStartE2EDuration="12.340849898s" podCreationTimestamp="2025-11-27 11:57:58 +0000 UTC" firstStartedPulling="2025-11-27 11:58:04.909220028 +0000 UTC m=+2002.427538946" lastFinishedPulling="2025-11-27 11:58:09.180939975 +0000 UTC m=+2006.699258893" observedRunningTime="2025-11-27 11:58:10.322298823 +0000 UTC m=+2007.840617741" watchObservedRunningTime="2025-11-27 11:58:10.340849898 +0000 UTC m=+2007.859168816" Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.320683 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7681c6a-f516-4b70-848c-2a6ce2c2e087","Type":"ContainerStarted","Data":"93f6838517958c4db07bafda85e164663e2a8a8ee82dafdba0819a78009e4923"} Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.325472 4796 generic.go:334] "Generic (PLEG): container finished" podID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerID="cea66ec29760c6fcb188ebf6db18b5a1ed51b73cd112a5af78186041a7ed3f50" exitCode=0 Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.325741 4796 generic.go:334] "Generic (PLEG): container finished" podID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerID="96b61b12f6b0b5ac9895c143d42c70c802b66089984ee792461cac0c5a45d221" exitCode=2 Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.325835 4796 generic.go:334] "Generic (PLEG): container finished" podID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerID="4b596e5b2fa264d353b58566df2213d77eb6c2b768939ed737dfbbdf3fa43c8a" exitCode=0 Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.325526 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerDied","Data":"cea66ec29760c6fcb188ebf6db18b5a1ed51b73cd112a5af78186041a7ed3f50"} Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.325988 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerDied","Data":"96b61b12f6b0b5ac9895c143d42c70c802b66089984ee792461cac0c5a45d221"} Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.326008 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerDied","Data":"4b596e5b2fa264d353b58566df2213d77eb6c2b768939ed737dfbbdf3fa43c8a"} Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.326629 4796 scope.go:117] "RemoveContainer" containerID="6691b124c02b21d7d108860d721432fbf98c9815031c7b2ebb1a7f93017d0cb0" Nov 27 11:58:11 crc kubenswrapper[4796]: E1127 11:58:11.326841 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-f48d67bf7-xvkst_openstack(e3b2e688-5f6c-4ead-9296-88a305dca376)\"" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.326968 4796 scope.go:117] "RemoveContainer" containerID="f98b7a273c0acc71226485dbf4ac219d4ee6179d3ef20e09b2e8977c69c4cc80" Nov 27 11:58:11 crc kubenswrapper[4796]: E1127 11:58:11.327319 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6fc467f47d-q8ztt_openstack(8fd6de03-a43e-44b2-97d0-ac73cf877eeb)\"" pod="openstack/heat-api-6fc467f47d-q8ztt" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" Nov 27 11:58:11 crc kubenswrapper[4796]: I1127 11:58:11.340581 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.340563757 podStartE2EDuration="3.340563757s" podCreationTimestamp="2025-11-27 11:58:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:11.339660543 +0000 UTC m=+2008.857979461" watchObservedRunningTime="2025-11-27 11:58:11.340563757 +0000 UTC m=+2008.858882685" Nov 27 11:58:12 crc kubenswrapper[4796]: I1127 11:58:12.596994 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5b7ccbdf6f-g754q" Nov 27 11:58:12 crc kubenswrapper[4796]: I1127 11:58:12.654024 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f48d67bf7-xvkst"] Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.057051 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.167863 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whspp\" (UniqueName: \"kubernetes.io/projected/e3b2e688-5f6c-4ead-9296-88a305dca376-kube-api-access-whspp\") pod \"e3b2e688-5f6c-4ead-9296-88a305dca376\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.167952 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-combined-ca-bundle\") pod \"e3b2e688-5f6c-4ead-9296-88a305dca376\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.168001 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data-custom\") pod \"e3b2e688-5f6c-4ead-9296-88a305dca376\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.168028 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data\") pod \"e3b2e688-5f6c-4ead-9296-88a305dca376\" (UID: \"e3b2e688-5f6c-4ead-9296-88a305dca376\") " Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.174284 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e3b2e688-5f6c-4ead-9296-88a305dca376" (UID: "e3b2e688-5f6c-4ead-9296-88a305dca376"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.174502 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b2e688-5f6c-4ead-9296-88a305dca376-kube-api-access-whspp" (OuterVolumeSpecName: "kube-api-access-whspp") pod "e3b2e688-5f6c-4ead-9296-88a305dca376" (UID: "e3b2e688-5f6c-4ead-9296-88a305dca376"). InnerVolumeSpecName "kube-api-access-whspp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.200203 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3b2e688-5f6c-4ead-9296-88a305dca376" (UID: "e3b2e688-5f6c-4ead-9296-88a305dca376"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.235450 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data" (OuterVolumeSpecName: "config-data") pod "e3b2e688-5f6c-4ead-9296-88a305dca376" (UID: "e3b2e688-5f6c-4ead-9296-88a305dca376"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.270708 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whspp\" (UniqueName: \"kubernetes.io/projected/e3b2e688-5f6c-4ead-9296-88a305dca376-kube-api-access-whspp\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.270747 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.270760 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.270771 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b2e688-5f6c-4ead-9296-88a305dca376-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.345441 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" event={"ID":"e3b2e688-5f6c-4ead-9296-88a305dca376","Type":"ContainerDied","Data":"1dba3b6bfda18e171d37e8415d52456542a5e0982fc856bf91e689bd0caa23ee"} Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.345498 4796 scope.go:117] "RemoveContainer" containerID="6691b124c02b21d7d108860d721432fbf98c9815031c7b2ebb1a7f93017d0cb0" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.345604 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f48d67bf7-xvkst" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.379160 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f48d67bf7-xvkst"] Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.390801 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-f48d67bf7-xvkst"] Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.584777 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" path="/var/lib/kubelet/pods/e3b2e688-5f6c-4ead-9296-88a305dca376/volumes" Nov 27 11:58:13 crc kubenswrapper[4796]: I1127 11:58:13.658922 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:58:16 crc kubenswrapper[4796]: I1127 11:58:16.714303 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 11:58:16 crc kubenswrapper[4796]: I1127 11:58:16.715739 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 27 11:58:16 crc kubenswrapper[4796]: I1127 11:58:16.760560 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 11:58:16 crc kubenswrapper[4796]: I1127 11:58:16.770697 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 27 11:58:17 crc kubenswrapper[4796]: I1127 11:58:17.390033 4796 generic.go:334] "Generic (PLEG): container finished" podID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerID="593b8a46f9c3ec5377b6dcfa338c3e2fe91a8b9e4e881611b0f04ccfa06097be" exitCode=0 Nov 27 11:58:17 crc kubenswrapper[4796]: I1127 11:58:17.390125 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerDied","Data":"593b8a46f9c3ec5377b6dcfa338c3e2fe91a8b9e4e881611b0f04ccfa06097be"} Nov 27 11:58:17 crc kubenswrapper[4796]: I1127 11:58:17.390551 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 11:58:17 crc kubenswrapper[4796]: I1127 11:58:17.390581 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 27 11:58:17 crc kubenswrapper[4796]: I1127 11:58:17.769434 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6896844f76-vnbq9" Nov 27 11:58:17 crc kubenswrapper[4796]: I1127 11:58:17.826319 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6fc467f47d-q8ztt"] Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.363875 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.368965 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.423669 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fc467f47d-q8ztt" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.423898 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fc467f47d-q8ztt" event={"ID":"8fd6de03-a43e-44b2-97d0-ac73cf877eeb","Type":"ContainerDied","Data":"d729e7755481800c9fc46f6cba308b6c12752390a7d490b22fb37af9079474a8"} Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.423951 4796 scope.go:117] "RemoveContainer" containerID="f98b7a273c0acc71226485dbf4ac219d4ee6179d3ef20e09b2e8977c69c4cc80" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.439026 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"452c5cce-4b40-4f0b-8975-4c395f8a703f","Type":"ContainerDied","Data":"10ed894bf6adb736717458dccb31bfb5e15a5cc36dbec5c5f329cc14ab0dc3bf"} Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.439211 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465531 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data-custom\") pod \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465608 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-config-data\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465639 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-sg-core-conf-yaml\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465678 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-scripts\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465719 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data\") pod \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465782 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-run-httpd\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465804 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-log-httpd\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465842 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-combined-ca-bundle\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465911 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lgh8\" (UniqueName: \"kubernetes.io/projected/452c5cce-4b40-4f0b-8975-4c395f8a703f-kube-api-access-4lgh8\") pod \"452c5cce-4b40-4f0b-8975-4c395f8a703f\" (UID: \"452c5cce-4b40-4f0b-8975-4c395f8a703f\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465958 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-combined-ca-bundle\") pod \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.465998 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n49th\" (UniqueName: \"kubernetes.io/projected/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-kube-api-access-n49th\") pod \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\" (UID: \"8fd6de03-a43e-44b2-97d0-ac73cf877eeb\") " Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.466822 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.468426 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.474879 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-kube-api-access-n49th" (OuterVolumeSpecName: "kube-api-access-n49th") pod "8fd6de03-a43e-44b2-97d0-ac73cf877eeb" (UID: "8fd6de03-a43e-44b2-97d0-ac73cf877eeb"). InnerVolumeSpecName "kube-api-access-n49th". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.475296 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8fd6de03-a43e-44b2-97d0-ac73cf877eeb" (UID: "8fd6de03-a43e-44b2-97d0-ac73cf877eeb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.496215 4796 scope.go:117] "RemoveContainer" containerID="cea66ec29760c6fcb188ebf6db18b5a1ed51b73cd112a5af78186041a7ed3f50" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.497171 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452c5cce-4b40-4f0b-8975-4c395f8a703f-kube-api-access-4lgh8" (OuterVolumeSpecName: "kube-api-access-4lgh8") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "kube-api-access-4lgh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.514509 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-scripts" (OuterVolumeSpecName: "scripts") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.543798 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.568843 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.569944 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/452c5cce-4b40-4f0b-8975-4c395f8a703f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.569981 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lgh8\" (UniqueName: \"kubernetes.io/projected/452c5cce-4b40-4f0b-8975-4c395f8a703f-kube-api-access-4lgh8\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.569995 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n49th\" (UniqueName: \"kubernetes.io/projected/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-kube-api-access-n49th\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.570006 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.570015 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.570024 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.571221 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data" (OuterVolumeSpecName: "config-data") pod "8fd6de03-a43e-44b2-97d0-ac73cf877eeb" (UID: "8fd6de03-a43e-44b2-97d0-ac73cf877eeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.576073 4796 scope.go:117] "RemoveContainer" containerID="96b61b12f6b0b5ac9895c143d42c70c802b66089984ee792461cac0c5a45d221" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.578626 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fd6de03-a43e-44b2-97d0-ac73cf877eeb" (UID: "8fd6de03-a43e-44b2-97d0-ac73cf877eeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.589423 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.605291 4796 scope.go:117] "RemoveContainer" containerID="4b596e5b2fa264d353b58566df2213d77eb6c2b768939ed737dfbbdf3fa43c8a" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.611600 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-config-data" (OuterVolumeSpecName: "config-data") pod "452c5cce-4b40-4f0b-8975-4c395f8a703f" (UID: "452c5cce-4b40-4f0b-8975-4c395f8a703f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.627445 4796 scope.go:117] "RemoveContainer" containerID="593b8a46f9c3ec5377b6dcfa338c3e2fe91a8b9e4e881611b0f04ccfa06097be" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.672032 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.672064 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.672080 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452c5cce-4b40-4f0b-8975-4c395f8a703f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.672093 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fd6de03-a43e-44b2-97d0-ac73cf877eeb-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.756420 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6fc467f47d-q8ztt"] Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.766702 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6fc467f47d-q8ztt"] Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.781107 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.806707 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.824467 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.824948 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerName="heat-cfnapi" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.824965 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerName="heat-cfnapi" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.824976 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-notification-agent" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.824983 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-notification-agent" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.825004 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="sg-core" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825012 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="sg-core" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.825025 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerName="heat-api" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825030 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerName="heat-api" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.825038 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerName="heat-api" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825044 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerName="heat-api" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.825050 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-central-agent" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825056 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-central-agent" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.825086 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="proxy-httpd" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825092 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="proxy-httpd" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825261 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-notification-agent" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825308 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerName="heat-api" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825333 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="proxy-httpd" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825346 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerName="heat-cfnapi" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825356 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="sg-core" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825367 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" containerName="ceilometer-central-agent" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825378 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerName="heat-cfnapi" Nov 27 11:58:18 crc kubenswrapper[4796]: E1127 11:58:18.825550 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerName="heat-cfnapi" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825559 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b2e688-5f6c-4ead-9296-88a305dca376" containerName="heat-cfnapi" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.825721 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" containerName="heat-api" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.827063 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.829297 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.829769 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.832975 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.845412 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.845603 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.875738 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-config-data\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.875830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.875884 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-log-httpd\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.875907 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tdsm\" (UniqueName: \"kubernetes.io/projected/33824a00-ea33-4982-88d5-a67b9b47f553-kube-api-access-8tdsm\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.875932 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.875964 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-scripts\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.876044 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-run-httpd\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.882958 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.894228 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977564 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-config-data\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977648 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977696 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-log-httpd\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977718 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tdsm\" (UniqueName: \"kubernetes.io/projected/33824a00-ea33-4982-88d5-a67b9b47f553-kube-api-access-8tdsm\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977737 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977755 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-scripts\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.977818 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-run-httpd\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.978429 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-log-httpd\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.978449 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-run-httpd\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.981642 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.981642 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.984575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-scripts\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.985200 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-config-data\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:18 crc kubenswrapper[4796]: I1127 11:58:18.996053 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tdsm\" (UniqueName: \"kubernetes.io/projected/33824a00-ea33-4982-88d5-a67b9b47f553-kube-api-access-8tdsm\") pod \"ceilometer-0\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " pod="openstack/ceilometer-0" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.144007 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.464982 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.466154 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.569510 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:58:19 crc kubenswrapper[4796]: E1127 11:58:19.569777 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.580660 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452c5cce-4b40-4f0b-8975-4c395f8a703f" path="/var/lib/kubelet/pods/452c5cce-4b40-4f0b-8975-4c395f8a703f/volumes" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.581595 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd6de03-a43e-44b2-97d0-ac73cf877eeb" path="/var/lib/kubelet/pods/8fd6de03-a43e-44b2-97d0-ac73cf877eeb/volumes" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.582191 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5469764df4-8wt46" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.637985 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.649808 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-658b556c84-2wb2w"] Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.650074 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-658b556c84-2wb2w" podUID="354047e3-61d2-4505-93f6-86592d3f62a7" containerName="heat-engine" containerID="cri-o://cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" gracePeriod=60 Nov 27 11:58:19 crc kubenswrapper[4796]: W1127 11:58:19.666845 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33824a00_ea33_4982_88d5_a67b9b47f553.slice/crio-15096684b33811f30ff0460f5eb261db8463db8533352cc715b4794b57dc6d85 WatchSource:0}: Error finding container 15096684b33811f30ff0460f5eb261db8463db8533352cc715b4794b57dc6d85: Status 404 returned error can't find the container with id 15096684b33811f30ff0460f5eb261db8463db8533352cc715b4794b57dc6d85 Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.846499 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.846606 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:58:19 crc kubenswrapper[4796]: I1127 11:58:19.899724 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 27 11:58:20 crc kubenswrapper[4796]: I1127 11:58:20.473803 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerStarted","Data":"15096684b33811f30ff0460f5eb261db8463db8533352cc715b4794b57dc6d85"} Nov 27 11:58:21 crc kubenswrapper[4796]: I1127 11:58:21.506476 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:58:21 crc kubenswrapper[4796]: I1127 11:58:21.506564 4796 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 11:58:21 crc kubenswrapper[4796]: I1127 11:58:21.870983 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:21 crc kubenswrapper[4796]: I1127 11:58:21.873322 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 27 11:58:22 crc kubenswrapper[4796]: I1127 11:58:22.515617 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerStarted","Data":"55e6875f02b857aaeacbc60b64d00434e72f6792cc7951ca3695f295919c8fdb"} Nov 27 11:58:23 crc kubenswrapper[4796]: I1127 11:58:23.483074 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:23 crc kubenswrapper[4796]: I1127 11:58:23.561752 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerStarted","Data":"8b0f7cedb6e09299c55c81ef526797a5b8e5e26f634989a3a40dd3cfe7d47761"} Nov 27 11:58:23 crc kubenswrapper[4796]: I1127 11:58:23.573328 4796 generic.go:334] "Generic (PLEG): container finished" podID="6524424e-afc2-443c-beac-c0503ac10ddf" containerID="064006ab2a9695d2a447ea377d0daf5fc0c97676ebc78389fc16ebd4dd27c40d" exitCode=0 Nov 27 11:58:23 crc kubenswrapper[4796]: I1127 11:58:23.594719 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-whbbq" event={"ID":"6524424e-afc2-443c-beac-c0503ac10ddf","Type":"ContainerDied","Data":"064006ab2a9695d2a447ea377d0daf5fc0c97676ebc78389fc16ebd4dd27c40d"} Nov 27 11:58:23 crc kubenswrapper[4796]: E1127 11:58:23.634613 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6524424e_afc2_443c_beac_c0503ac10ddf.slice/crio-conmon-064006ab2a9695d2a447ea377d0daf5fc0c97676ebc78389fc16ebd4dd27c40d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6524424e_afc2_443c_beac_c0503ac10ddf.slice/crio-064006ab2a9695d2a447ea377d0daf5fc0c97676ebc78389fc16ebd4dd27c40d.scope\": RecentStats: unable to find data in memory cache]" Nov 27 11:58:23 crc kubenswrapper[4796]: E1127 11:58:23.640547 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 27 11:58:23 crc kubenswrapper[4796]: E1127 11:58:23.649681 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 27 11:58:23 crc kubenswrapper[4796]: E1127 11:58:23.665552 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 27 11:58:23 crc kubenswrapper[4796]: E1127 11:58:23.665624 4796 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-658b556c84-2wb2w" podUID="354047e3-61d2-4505-93f6-86592d3f62a7" containerName="heat-engine" Nov 27 11:58:24 crc kubenswrapper[4796]: I1127 11:58:24.587105 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerStarted","Data":"7eadc11b12c4e52595f07d3f39ce7316dddd6844ee33a695ed64486343577fb4"} Nov 27 11:58:24 crc kubenswrapper[4796]: I1127 11:58:24.959464 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.011929 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-scripts\") pod \"6524424e-afc2-443c-beac-c0503ac10ddf\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.012150 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgf82\" (UniqueName: \"kubernetes.io/projected/6524424e-afc2-443c-beac-c0503ac10ddf-kube-api-access-kgf82\") pod \"6524424e-afc2-443c-beac-c0503ac10ddf\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.012186 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-combined-ca-bundle\") pod \"6524424e-afc2-443c-beac-c0503ac10ddf\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.012246 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-config-data\") pod \"6524424e-afc2-443c-beac-c0503ac10ddf\" (UID: \"6524424e-afc2-443c-beac-c0503ac10ddf\") " Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.018880 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6524424e-afc2-443c-beac-c0503ac10ddf-kube-api-access-kgf82" (OuterVolumeSpecName: "kube-api-access-kgf82") pod "6524424e-afc2-443c-beac-c0503ac10ddf" (UID: "6524424e-afc2-443c-beac-c0503ac10ddf"). InnerVolumeSpecName "kube-api-access-kgf82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.022303 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-scripts" (OuterVolumeSpecName: "scripts") pod "6524424e-afc2-443c-beac-c0503ac10ddf" (UID: "6524424e-afc2-443c-beac-c0503ac10ddf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.042742 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-config-data" (OuterVolumeSpecName: "config-data") pod "6524424e-afc2-443c-beac-c0503ac10ddf" (UID: "6524424e-afc2-443c-beac-c0503ac10ddf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.045961 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6524424e-afc2-443c-beac-c0503ac10ddf" (UID: "6524424e-afc2-443c-beac-c0503ac10ddf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.114816 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgf82\" (UniqueName: \"kubernetes.io/projected/6524424e-afc2-443c-beac-c0503ac10ddf-kube-api-access-kgf82\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.114861 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.114873 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.114887 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6524424e-afc2-443c-beac-c0503ac10ddf-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.622040 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-whbbq" event={"ID":"6524424e-afc2-443c-beac-c0503ac10ddf","Type":"ContainerDied","Data":"c11ca53fce62e703be022b1b74a6bfff1a7626a983ddb45edbe457cb02313231"} Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.622335 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c11ca53fce62e703be022b1b74a6bfff1a7626a983ddb45edbe457cb02313231" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.622401 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-whbbq" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.754364 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 11:58:25 crc kubenswrapper[4796]: E1127 11:58:25.754802 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6524424e-afc2-443c-beac-c0503ac10ddf" containerName="nova-cell0-conductor-db-sync" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.754820 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6524424e-afc2-443c-beac-c0503ac10ddf" containerName="nova-cell0-conductor-db-sync" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.755006 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6524424e-afc2-443c-beac-c0503ac10ddf" containerName="nova-cell0-conductor-db-sync" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.755622 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.760509 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.760769 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4htnn" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.769851 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.834032 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-579dz\" (UniqueName: \"kubernetes.io/projected/a30067ee-e2bf-4d5f-bdcf-cff461743b62-kube-api-access-579dz\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.834087 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30067ee-e2bf-4d5f-bdcf-cff461743b62-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.834422 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30067ee-e2bf-4d5f-bdcf-cff461743b62-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.936813 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-579dz\" (UniqueName: \"kubernetes.io/projected/a30067ee-e2bf-4d5f-bdcf-cff461743b62-kube-api-access-579dz\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.936872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30067ee-e2bf-4d5f-bdcf-cff461743b62-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.936950 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30067ee-e2bf-4d5f-bdcf-cff461743b62-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.942684 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30067ee-e2bf-4d5f-bdcf-cff461743b62-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.945096 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30067ee-e2bf-4d5f-bdcf-cff461743b62-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:25 crc kubenswrapper[4796]: I1127 11:58:25.960146 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-579dz\" (UniqueName: \"kubernetes.io/projected/a30067ee-e2bf-4d5f-bdcf-cff461743b62-kube-api-access-579dz\") pod \"nova-cell0-conductor-0\" (UID: \"a30067ee-e2bf-4d5f-bdcf-cff461743b62\") " pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.075085 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.633322 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerStarted","Data":"6c4a3c09bed6841db174581198276b661adce6c01ceb6d8ae094cca4331fde06"} Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.633913 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.633551 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="proxy-httpd" containerID="cri-o://6c4a3c09bed6841db174581198276b661adce6c01ceb6d8ae094cca4331fde06" gracePeriod=30 Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.633442 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-central-agent" containerID="cri-o://55e6875f02b857aaeacbc60b64d00434e72f6792cc7951ca3695f295919c8fdb" gracePeriod=30 Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.633573 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="sg-core" containerID="cri-o://7eadc11b12c4e52595f07d3f39ce7316dddd6844ee33a695ed64486343577fb4" gracePeriod=30 Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.633566 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-notification-agent" containerID="cri-o://8b0f7cedb6e09299c55c81ef526797a5b8e5e26f634989a3a40dd3cfe7d47761" gracePeriod=30 Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.667755 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.29428233 podStartE2EDuration="8.667734536s" podCreationTimestamp="2025-11-27 11:58:18 +0000 UTC" firstStartedPulling="2025-11-27 11:58:19.670407165 +0000 UTC m=+2017.188726083" lastFinishedPulling="2025-11-27 11:58:26.043859371 +0000 UTC m=+2023.562178289" observedRunningTime="2025-11-27 11:58:26.650667351 +0000 UTC m=+2024.168986269" watchObservedRunningTime="2025-11-27 11:58:26.667734536 +0000 UTC m=+2024.186053454" Nov 27 11:58:26 crc kubenswrapper[4796]: I1127 11:58:26.688324 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 11:58:26 crc kubenswrapper[4796]: W1127 11:58:26.693805 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda30067ee_e2bf_4d5f_bdcf_cff461743b62.slice/crio-832236f2b4739e82df1cd0309ca79c60f610eff7bd3b06c70776c6c46f5d3386 WatchSource:0}: Error finding container 832236f2b4739e82df1cd0309ca79c60f610eff7bd3b06c70776c6c46f5d3386: Status 404 returned error can't find the container with id 832236f2b4739e82df1cd0309ca79c60f610eff7bd3b06c70776c6c46f5d3386 Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.281443 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.364449 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-combined-ca-bundle\") pod \"354047e3-61d2-4505-93f6-86592d3f62a7\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.364527 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l7nb\" (UniqueName: \"kubernetes.io/projected/354047e3-61d2-4505-93f6-86592d3f62a7-kube-api-access-4l7nb\") pod \"354047e3-61d2-4505-93f6-86592d3f62a7\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.364656 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data-custom\") pod \"354047e3-61d2-4505-93f6-86592d3f62a7\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.364741 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data\") pod \"354047e3-61d2-4505-93f6-86592d3f62a7\" (UID: \"354047e3-61d2-4505-93f6-86592d3f62a7\") " Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.379452 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354047e3-61d2-4505-93f6-86592d3f62a7-kube-api-access-4l7nb" (OuterVolumeSpecName: "kube-api-access-4l7nb") pod "354047e3-61d2-4505-93f6-86592d3f62a7" (UID: "354047e3-61d2-4505-93f6-86592d3f62a7"). InnerVolumeSpecName "kube-api-access-4l7nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.398404 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "354047e3-61d2-4505-93f6-86592d3f62a7" (UID: "354047e3-61d2-4505-93f6-86592d3f62a7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.432448 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "354047e3-61d2-4505-93f6-86592d3f62a7" (UID: "354047e3-61d2-4505-93f6-86592d3f62a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.466521 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.466553 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l7nb\" (UniqueName: \"kubernetes.io/projected/354047e3-61d2-4505-93f6-86592d3f62a7-kube-api-access-4l7nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.466563 4796 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.476437 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data" (OuterVolumeSpecName: "config-data") pod "354047e3-61d2-4505-93f6-86592d3f62a7" (UID: "354047e3-61d2-4505-93f6-86592d3f62a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.568071 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354047e3-61d2-4505-93f6-86592d3f62a7-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.656153 4796 generic.go:334] "Generic (PLEG): container finished" podID="33824a00-ea33-4982-88d5-a67b9b47f553" containerID="6c4a3c09bed6841db174581198276b661adce6c01ceb6d8ae094cca4331fde06" exitCode=0 Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.657012 4796 generic.go:334] "Generic (PLEG): container finished" podID="33824a00-ea33-4982-88d5-a67b9b47f553" containerID="7eadc11b12c4e52595f07d3f39ce7316dddd6844ee33a695ed64486343577fb4" exitCode=2 Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.657085 4796 generic.go:334] "Generic (PLEG): container finished" podID="33824a00-ea33-4982-88d5-a67b9b47f553" containerID="8b0f7cedb6e09299c55c81ef526797a5b8e5e26f634989a3a40dd3cfe7d47761" exitCode=0 Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.656305 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerDied","Data":"6c4a3c09bed6841db174581198276b661adce6c01ceb6d8ae094cca4331fde06"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.657548 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerDied","Data":"7eadc11b12c4e52595f07d3f39ce7316dddd6844ee33a695ed64486343577fb4"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.657617 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerDied","Data":"8b0f7cedb6e09299c55c81ef526797a5b8e5e26f634989a3a40dd3cfe7d47761"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.661769 4796 generic.go:334] "Generic (PLEG): container finished" podID="354047e3-61d2-4505-93f6-86592d3f62a7" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" exitCode=0 Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.661838 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-658b556c84-2wb2w" event={"ID":"354047e3-61d2-4505-93f6-86592d3f62a7","Type":"ContainerDied","Data":"cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.661866 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-658b556c84-2wb2w" event={"ID":"354047e3-61d2-4505-93f6-86592d3f62a7","Type":"ContainerDied","Data":"e18339c50968a657c05bd92131a3e2a2a1fa9eb2d46de45758659310eb93339f"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.661881 4796 scope.go:117] "RemoveContainer" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.662002 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-658b556c84-2wb2w" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.675608 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a30067ee-e2bf-4d5f-bdcf-cff461743b62","Type":"ContainerStarted","Data":"6b86492fb18f76cda64fd187ab97bb4ef33ae9d4bd51cc3195f80005550bd4c8"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.675709 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a30067ee-e2bf-4d5f-bdcf-cff461743b62","Type":"ContainerStarted","Data":"832236f2b4739e82df1cd0309ca79c60f610eff7bd3b06c70776c6c46f5d3386"} Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.675949 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.687557 4796 scope.go:117] "RemoveContainer" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" Nov 27 11:58:27 crc kubenswrapper[4796]: E1127 11:58:27.689024 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21\": container with ID starting with cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21 not found: ID does not exist" containerID="cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.689073 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21"} err="failed to get container status \"cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21\": rpc error: code = NotFound desc = could not find container \"cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21\": container with ID starting with cbc266cb07003bffbca2ecc5d3dc916331333ab220f5d08f399f91f29eb16e21 not found: ID does not exist" Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.703351 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-658b556c84-2wb2w"] Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.727759 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-658b556c84-2wb2w"] Nov 27 11:58:27 crc kubenswrapper[4796]: I1127 11:58:27.728838 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.728817549 podStartE2EDuration="2.728817549s" podCreationTimestamp="2025-11-27 11:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:27.701588907 +0000 UTC m=+2025.219907835" watchObservedRunningTime="2025-11-27 11:58:27.728817549 +0000 UTC m=+2025.247136467" Nov 27 11:58:29 crc kubenswrapper[4796]: I1127 11:58:29.579494 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="354047e3-61d2-4505-93f6-86592d3f62a7" path="/var/lib/kubelet/pods/354047e3-61d2-4505-93f6-86592d3f62a7/volumes" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.134994 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.569822 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:58:31 crc kubenswrapper[4796]: E1127 11:58:31.570125 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.702965 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lbg4j"] Nov 27 11:58:31 crc kubenswrapper[4796]: E1127 11:58:31.703461 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354047e3-61d2-4505-93f6-86592d3f62a7" containerName="heat-engine" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.703485 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="354047e3-61d2-4505-93f6-86592d3f62a7" containerName="heat-engine" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.703722 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="354047e3-61d2-4505-93f6-86592d3f62a7" containerName="heat-engine" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.704517 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.725431 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.729770 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.742756 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.742826 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-scripts\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.742935 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.742977 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjhj\" (UniqueName: \"kubernetes.io/projected/9cbdc050-92dd-44be-83f2-474365084f53-kube-api-access-4jjhj\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.789214 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lbg4j"] Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.849232 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.849309 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjhj\" (UniqueName: \"kubernetes.io/projected/9cbdc050-92dd-44be-83f2-474365084f53-kube-api-access-4jjhj\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.849339 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.849375 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-scripts\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.865074 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.865815 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-scripts\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.882447 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.883756 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjhj\" (UniqueName: \"kubernetes.io/projected/9cbdc050-92dd-44be-83f2-474365084f53-kube-api-access-4jjhj\") pod \"nova-cell0-cell-mapping-lbg4j\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.989171 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:31 crc kubenswrapper[4796]: I1127 11:58:31.990667 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.041686 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.043554 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.058140 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg8tg\" (UniqueName: \"kubernetes.io/projected/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-kube-api-access-xg8tg\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.058251 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.058338 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.058368 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-logs\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.068010 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.069216 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.078683 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.091357 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.101004 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161487 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161572 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161610 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-logs\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161650 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161681 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8tg\" (UniqueName: \"kubernetes.io/projected/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-kube-api-access-xg8tg\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161704 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fqdt\" (UniqueName: \"kubernetes.io/projected/69926936-b4f5-4201-8df2-3ac92004ad81-kube-api-access-4fqdt\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.161751 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-config-data\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.166347 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-logs\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.171652 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.171755 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.174216 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.179469 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.186412 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.186630 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.206863 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg8tg\" (UniqueName: \"kubernetes.io/projected/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-kube-api-access-xg8tg\") pod \"nova-api-0\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.221764 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.223199 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.227359 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269608 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269672 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a4b071-b7f0-4b74-8ccb-740eef31f462-logs\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269695 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269716 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fqdt\" (UniqueName: \"kubernetes.io/projected/69926936-b4f5-4201-8df2-3ac92004ad81-kube-api-access-4fqdt\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269738 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrrx7\" (UniqueName: \"kubernetes.io/projected/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-kube-api-access-vrrx7\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269767 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-config-data\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269795 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-config-data\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269810 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269845 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bg7h\" (UniqueName: \"kubernetes.io/projected/21a4b071-b7f0-4b74-8ccb-740eef31f462-kube-api-access-2bg7h\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.269873 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.279449 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-config-data\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.282223 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.307761 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fqdt\" (UniqueName: \"kubernetes.io/projected/69926936-b4f5-4201-8df2-3ac92004ad81-kube-api-access-4fqdt\") pod \"nova-scheduler-0\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.314463 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.317210 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371531 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371650 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a4b071-b7f0-4b74-8ccb-740eef31f462-logs\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371670 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371690 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrrx7\" (UniqueName: \"kubernetes.io/projected/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-kube-api-access-vrrx7\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371718 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-config-data\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.371775 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bg7h\" (UniqueName: \"kubernetes.io/projected/21a4b071-b7f0-4b74-8ccb-740eef31f462-kube-api-access-2bg7h\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.378972 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-kg5xv"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.381919 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a4b071-b7f0-4b74-8ccb-740eef31f462-logs\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.382603 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.383832 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.383881 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-config-data\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.387660 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.390132 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.390577 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bg7h\" (UniqueName: \"kubernetes.io/projected/21a4b071-b7f0-4b74-8ccb-740eef31f462-kube-api-access-2bg7h\") pod \"nova-metadata-0\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.395245 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrrx7\" (UniqueName: \"kubernetes.io/projected/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-kube-api-access-vrrx7\") pod \"nova-cell1-novncproxy-0\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.417584 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-kg5xv"] Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.475715 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnsqq\" (UniqueName: \"kubernetes.io/projected/e17ae35d-43d7-4afb-947e-db43516a6a92-kube-api-access-xnsqq\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.476146 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-config\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.476230 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.476486 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.476585 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.476611 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.549729 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.579040 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-config\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.579112 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.579180 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.579216 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.579248 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.579335 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnsqq\" (UniqueName: \"kubernetes.io/projected/e17ae35d-43d7-4afb-947e-db43516a6a92-kube-api-access-xnsqq\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.581128 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-config\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.581958 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.582661 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.583210 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.583767 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.596608 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnsqq\" (UniqueName: \"kubernetes.io/projected/e17ae35d-43d7-4afb-947e-db43516a6a92-kube-api-access-xnsqq\") pod \"dnsmasq-dns-568d7fd7cf-kg5xv\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.602356 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.631595 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.726896 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.803187 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lbg4j"] Nov 27 11:58:32 crc kubenswrapper[4796]: W1127 11:58:32.812470 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cbdc050_92dd_44be_83f2_474365084f53.slice/crio-45a34489e8ea5cb39113a2bc12c0762842258297a4b558766d022cbbafbf07c2 WatchSource:0}: Error finding container 45a34489e8ea5cb39113a2bc12c0762842258297a4b558766d022cbbafbf07c2: Status 404 returned error can't find the container with id 45a34489e8ea5cb39113a2bc12c0762842258297a4b558766d022cbbafbf07c2 Nov 27 11:58:32 crc kubenswrapper[4796]: I1127 11:58:32.989404 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.282148 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.385290 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.403137 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-kg5xv"] Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.566498 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.742925 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69926936-b4f5-4201-8df2-3ac92004ad81","Type":"ContainerStarted","Data":"e9b21e9bd04e9994f2d83421b0d2dcd017afb17dcf8bc152225ce11106513f8d"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.750897 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85","Type":"ContainerStarted","Data":"fe5ffeecb664613185ab88f85f6ebd77a72be342a8235ea5ba6f6fd4bdebfdff"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.755518 4796 generic.go:334] "Generic (PLEG): container finished" podID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerID="04ed696bf6ee51aec2ae57cb48a8d6f078a90330b94ad247b1e0c165c899004e" exitCode=0 Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.755630 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" event={"ID":"e17ae35d-43d7-4afb-947e-db43516a6a92","Type":"ContainerDied","Data":"04ed696bf6ee51aec2ae57cb48a8d6f078a90330b94ad247b1e0c165c899004e"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.755659 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" event={"ID":"e17ae35d-43d7-4afb-947e-db43516a6a92","Type":"ContainerStarted","Data":"b436f7d479cee8ea816572ad754051293ee4f59aff87723b2e054dddedb0c90c"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.759072 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lbg4j" event={"ID":"9cbdc050-92dd-44be-83f2-474365084f53","Type":"ContainerStarted","Data":"eac3bfb17e80603b35583d733348e26451d16279d1d6bd997f982c96b2f362a6"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.759188 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lbg4j" event={"ID":"9cbdc050-92dd-44be-83f2-474365084f53","Type":"ContainerStarted","Data":"45a34489e8ea5cb39113a2bc12c0762842258297a4b558766d022cbbafbf07c2"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.760428 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21a4b071-b7f0-4b74-8ccb-740eef31f462","Type":"ContainerStarted","Data":"8b437154174b16d8abc631d94aecd568c27b1810196feb017ae94f979af754e3"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.761477 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c","Type":"ContainerStarted","Data":"f8d514fbff7adc2b770468c973bc81374e33a05fc6df86f0cd2084858620afc9"} Nov 27 11:58:33 crc kubenswrapper[4796]: I1127 11:58:33.795244 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lbg4j" podStartSLOduration=2.795226455 podStartE2EDuration="2.795226455s" podCreationTimestamp="2025-11-27 11:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:33.791187604 +0000 UTC m=+2031.309506512" watchObservedRunningTime="2025-11-27 11:58:33.795226455 +0000 UTC m=+2031.313545373" Nov 27 11:58:34 crc kubenswrapper[4796]: E1127 11:58:34.004368 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode17ae35d_43d7_4afb_947e_db43516a6a92.slice/crio-conmon-04ed696bf6ee51aec2ae57cb48a8d6f078a90330b94ad247b1e0c165c899004e.scope\": RecentStats: unable to find data in memory cache]" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.128758 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nzgfc"] Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.130006 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.133239 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.133436 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.143482 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nzgfc"] Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.234543 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-scripts\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.234626 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.234705 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwmlq\" (UniqueName: \"kubernetes.io/projected/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-kube-api-access-jwmlq\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.234789 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-config-data\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.336410 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-scripts\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.337499 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.337944 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwmlq\" (UniqueName: \"kubernetes.io/projected/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-kube-api-access-jwmlq\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.338110 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-config-data\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.377171 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.377794 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-scripts\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.377971 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-config-data\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.390774 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwmlq\" (UniqueName: \"kubernetes.io/projected/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-kube-api-access-jwmlq\") pod \"nova-cell1-conductor-db-sync-nzgfc\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.444459 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.801937 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" event={"ID":"e17ae35d-43d7-4afb-947e-db43516a6a92","Type":"ContainerStarted","Data":"7bf76cf318fe6e654a801a4bf85ab7f81f2e81f0051a883e87825affca93c2ca"} Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.804616 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:34 crc kubenswrapper[4796]: I1127 11:58:34.832579 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" podStartSLOduration=2.8325560899999997 podStartE2EDuration="2.83255609s" podCreationTimestamp="2025-11-27 11:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:34.822065103 +0000 UTC m=+2032.340384031" watchObservedRunningTime="2025-11-27 11:58:34.83255609 +0000 UTC m=+2032.350875008" Nov 27 11:58:35 crc kubenswrapper[4796]: I1127 11:58:35.079805 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nzgfc"] Nov 27 11:58:35 crc kubenswrapper[4796]: W1127 11:58:35.082939 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa3bdadd_98ed_4f1e_88c2_d23354a0e84a.slice/crio-c0f6fe650d76ea1283fe0b6ecc25959b4f85bb258315cc7fc3eabdb68fe55388 WatchSource:0}: Error finding container c0f6fe650d76ea1283fe0b6ecc25959b4f85bb258315cc7fc3eabdb68fe55388: Status 404 returned error can't find the container with id c0f6fe650d76ea1283fe0b6ecc25959b4f85bb258315cc7fc3eabdb68fe55388 Nov 27 11:58:35 crc kubenswrapper[4796]: I1127 11:58:35.825726 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" event={"ID":"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a","Type":"ContainerStarted","Data":"5593861da04285f8c0fd914fe3b491e2e50de627ed3718f78d306a76e8e08b28"} Nov 27 11:58:35 crc kubenswrapper[4796]: I1127 11:58:35.826003 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" event={"ID":"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a","Type":"ContainerStarted","Data":"c0f6fe650d76ea1283fe0b6ecc25959b4f85bb258315cc7fc3eabdb68fe55388"} Nov 27 11:58:35 crc kubenswrapper[4796]: I1127 11:58:35.852003 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" podStartSLOduration=1.851985187 podStartE2EDuration="1.851985187s" podCreationTimestamp="2025-11-27 11:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:35.847382621 +0000 UTC m=+2033.365701539" watchObservedRunningTime="2025-11-27 11:58:35.851985187 +0000 UTC m=+2033.370304105" Nov 27 11:58:35 crc kubenswrapper[4796]: I1127 11:58:35.914587 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:58:35 crc kubenswrapper[4796]: I1127 11:58:35.932624 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:36 crc kubenswrapper[4796]: I1127 11:58:36.836156 4796 generic.go:334] "Generic (PLEG): container finished" podID="33824a00-ea33-4982-88d5-a67b9b47f553" containerID="55e6875f02b857aaeacbc60b64d00434e72f6792cc7951ca3695f295919c8fdb" exitCode=0 Nov 27 11:58:36 crc kubenswrapper[4796]: I1127 11:58:36.836226 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerDied","Data":"55e6875f02b857aaeacbc60b64d00434e72f6792cc7951ca3695f295919c8fdb"} Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.784433 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.849383 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85","Type":"ContainerStarted","Data":"467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7"} Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.857730 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33824a00-ea33-4982-88d5-a67b9b47f553","Type":"ContainerDied","Data":"15096684b33811f30ff0460f5eb261db8463db8533352cc715b4794b57dc6d85"} Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.857770 4796 scope.go:117] "RemoveContainer" containerID="6c4a3c09bed6841db174581198276b661adce6c01ceb6d8ae094cca4331fde06" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.857906 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.860089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21a4b071-b7f0-4b74-8ccb-740eef31f462","Type":"ContainerStarted","Data":"5b27df5a83b802e33c8a4c14826b32e167bae1d47891b1f274714eed193ea090"} Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.861998 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c","Type":"ContainerStarted","Data":"46170f34192949854fbd2254a59b79816457c154546193bcc9a421c252b70e9f"} Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.862106 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://46170f34192949854fbd2254a59b79816457c154546193bcc9a421c252b70e9f" gracePeriod=30 Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.872434 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69926936-b4f5-4201-8df2-3ac92004ad81","Type":"ContainerStarted","Data":"db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519"} Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.879810 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.894447869 podStartE2EDuration="5.879791901s" podCreationTimestamp="2025-11-27 11:58:32 +0000 UTC" firstStartedPulling="2025-11-27 11:58:33.390392759 +0000 UTC m=+2030.908711677" lastFinishedPulling="2025-11-27 11:58:37.375736791 +0000 UTC m=+2034.894055709" observedRunningTime="2025-11-27 11:58:37.877003225 +0000 UTC m=+2035.395322143" watchObservedRunningTime="2025-11-27 11:58:37.879791901 +0000 UTC m=+2035.398110819" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.892438 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.796996832 podStartE2EDuration="6.892422974s" podCreationTimestamp="2025-11-27 11:58:31 +0000 UTC" firstStartedPulling="2025-11-27 11:58:33.299999996 +0000 UTC m=+2030.818318914" lastFinishedPulling="2025-11-27 11:58:37.395426138 +0000 UTC m=+2034.913745056" observedRunningTime="2025-11-27 11:58:37.890415419 +0000 UTC m=+2035.408734337" watchObservedRunningTime="2025-11-27 11:58:37.892422974 +0000 UTC m=+2035.410741892" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.893601 4796 scope.go:117] "RemoveContainer" containerID="7eadc11b12c4e52595f07d3f39ce7316dddd6844ee33a695ed64486343577fb4" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915620 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdsm\" (UniqueName: \"kubernetes.io/projected/33824a00-ea33-4982-88d5-a67b9b47f553-kube-api-access-8tdsm\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915714 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-run-httpd\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915760 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-combined-ca-bundle\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915859 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-config-data\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915889 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-log-httpd\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915912 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-sg-core-conf-yaml\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.915957 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-scripts\") pod \"33824a00-ea33-4982-88d5-a67b9b47f553\" (UID: \"33824a00-ea33-4982-88d5-a67b9b47f553\") " Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.918002 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.919490 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.920297 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-scripts" (OuterVolumeSpecName: "scripts") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.920941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33824a00-ea33-4982-88d5-a67b9b47f553-kube-api-access-8tdsm" (OuterVolumeSpecName: "kube-api-access-8tdsm") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "kube-api-access-8tdsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.931368 4796 scope.go:117] "RemoveContainer" containerID="8b0f7cedb6e09299c55c81ef526797a5b8e5e26f634989a3a40dd3cfe7d47761" Nov 27 11:58:37 crc kubenswrapper[4796]: I1127 11:58:37.958520 4796 scope.go:117] "RemoveContainer" containerID="55e6875f02b857aaeacbc60b64d00434e72f6792cc7951ca3695f295919c8fdb" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.018762 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdsm\" (UniqueName: \"kubernetes.io/projected/33824a00-ea33-4982-88d5-a67b9b47f553-kube-api-access-8tdsm\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.018791 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.018804 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33824a00-ea33-4982-88d5-a67b9b47f553-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.018815 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.028701 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.083357 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.120555 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.120583 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.123365 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-config-data" (OuterVolumeSpecName: "config-data") pod "33824a00-ea33-4982-88d5-a67b9b47f553" (UID: "33824a00-ea33-4982-88d5-a67b9b47f553"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.189940 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.197820 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.222462 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33824a00-ea33-4982-88d5-a67b9b47f553-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.222500 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:38 crc kubenswrapper[4796]: E1127 11:58:38.222886 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-central-agent" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.222902 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-central-agent" Nov 27 11:58:38 crc kubenswrapper[4796]: E1127 11:58:38.222938 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="sg-core" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.222946 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="sg-core" Nov 27 11:58:38 crc kubenswrapper[4796]: E1127 11:58:38.222955 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="proxy-httpd" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.222961 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="proxy-httpd" Nov 27 11:58:38 crc kubenswrapper[4796]: E1127 11:58:38.222975 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-notification-agent" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.222980 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-notification-agent" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.223156 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-notification-agent" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.223168 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="sg-core" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.223189 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="ceilometer-central-agent" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.223202 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" containerName="proxy-httpd" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.224769 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.236772 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.237170 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.280281 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324049 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8kbd\" (UniqueName: \"kubernetes.io/projected/522aafbf-5229-44e5-b61c-d194b5aca60b-kube-api-access-n8kbd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324108 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324150 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-run-httpd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324168 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324210 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-config-data\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324252 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-scripts\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.324291 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-log-httpd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426059 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-config-data\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426472 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-scripts\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426506 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-log-httpd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426553 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8kbd\" (UniqueName: \"kubernetes.io/projected/522aafbf-5229-44e5-b61c-d194b5aca60b-kube-api-access-n8kbd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426590 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426628 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-run-httpd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.426979 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-log-httpd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.427478 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-run-httpd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.432647 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.433129 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.433866 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-scripts\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.434197 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-config-data\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.449720 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8kbd\" (UniqueName: \"kubernetes.io/projected/522aafbf-5229-44e5-b61c-d194b5aca60b-kube-api-access-n8kbd\") pod \"ceilometer-0\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.543606 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.890113 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21a4b071-b7f0-4b74-8ccb-740eef31f462","Type":"ContainerStarted","Data":"4ba7d8e46301d427593616a3116554fb5ee8715265eee58d02e4204da21ec87a"} Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.890194 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-log" containerID="cri-o://5b27df5a83b802e33c8a4c14826b32e167bae1d47891b1f274714eed193ea090" gracePeriod=30 Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.890501 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-metadata" containerID="cri-o://4ba7d8e46301d427593616a3116554fb5ee8715265eee58d02e4204da21ec87a" gracePeriod=30 Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.905356 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85","Type":"ContainerStarted","Data":"781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f"} Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.930510 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.559590889 podStartE2EDuration="7.93049669s" podCreationTimestamp="2025-11-27 11:58:31 +0000 UTC" firstStartedPulling="2025-11-27 11:58:33.003524774 +0000 UTC m=+2030.521843692" lastFinishedPulling="2025-11-27 11:58:37.374430585 +0000 UTC m=+2034.892749493" observedRunningTime="2025-11-27 11:58:38.928663 +0000 UTC m=+2036.446981918" watchObservedRunningTime="2025-11-27 11:58:38.93049669 +0000 UTC m=+2036.448815608" Nov 27 11:58:38 crc kubenswrapper[4796]: I1127 11:58:38.931963 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.171649363 podStartE2EDuration="6.93195888s" podCreationTimestamp="2025-11-27 11:58:32 +0000 UTC" firstStartedPulling="2025-11-27 11:58:33.626100034 +0000 UTC m=+2031.144418952" lastFinishedPulling="2025-11-27 11:58:37.386409551 +0000 UTC m=+2034.904728469" observedRunningTime="2025-11-27 11:58:38.910436954 +0000 UTC m=+2036.428755872" watchObservedRunningTime="2025-11-27 11:58:38.93195888 +0000 UTC m=+2036.450277788" Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.026623 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:58:39 crc kubenswrapper[4796]: W1127 11:58:39.031163 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod522aafbf_5229_44e5_b61c_d194b5aca60b.slice/crio-64c9ee311f83df9fe0b71c1ba7ad9318d33773fc07ca4b183265d87e420bbeda WatchSource:0}: Error finding container 64c9ee311f83df9fe0b71c1ba7ad9318d33773fc07ca4b183265d87e420bbeda: Status 404 returned error can't find the container with id 64c9ee311f83df9fe0b71c1ba7ad9318d33773fc07ca4b183265d87e420bbeda Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.584890 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33824a00-ea33-4982-88d5-a67b9b47f553" path="/var/lib/kubelet/pods/33824a00-ea33-4982-88d5-a67b9b47f553/volumes" Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.916881 4796 generic.go:334] "Generic (PLEG): container finished" podID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerID="4ba7d8e46301d427593616a3116554fb5ee8715265eee58d02e4204da21ec87a" exitCode=0 Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.917170 4796 generic.go:334] "Generic (PLEG): container finished" podID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerID="5b27df5a83b802e33c8a4c14826b32e167bae1d47891b1f274714eed193ea090" exitCode=143 Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.917009 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21a4b071-b7f0-4b74-8ccb-740eef31f462","Type":"ContainerDied","Data":"4ba7d8e46301d427593616a3116554fb5ee8715265eee58d02e4204da21ec87a"} Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.917233 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21a4b071-b7f0-4b74-8ccb-740eef31f462","Type":"ContainerDied","Data":"5b27df5a83b802e33c8a4c14826b32e167bae1d47891b1f274714eed193ea090"} Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.920969 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerStarted","Data":"835a2be64574b9da32af7034626adf0469099c4ab8a94e726e11b866bc281b24"} Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.921020 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerStarted","Data":"64c9ee311f83df9fe0b71c1ba7ad9318d33773fc07ca4b183265d87e420bbeda"} Nov 27 11:58:39 crc kubenswrapper[4796]: I1127 11:58:39.997672 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.178282 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bg7h\" (UniqueName: \"kubernetes.io/projected/21a4b071-b7f0-4b74-8ccb-740eef31f462-kube-api-access-2bg7h\") pod \"21a4b071-b7f0-4b74-8ccb-740eef31f462\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.178633 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a4b071-b7f0-4b74-8ccb-740eef31f462-logs\") pod \"21a4b071-b7f0-4b74-8ccb-740eef31f462\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.178814 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-config-data\") pod \"21a4b071-b7f0-4b74-8ccb-740eef31f462\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.178974 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-combined-ca-bundle\") pod \"21a4b071-b7f0-4b74-8ccb-740eef31f462\" (UID: \"21a4b071-b7f0-4b74-8ccb-740eef31f462\") " Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.178866 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21a4b071-b7f0-4b74-8ccb-740eef31f462-logs" (OuterVolumeSpecName: "logs") pod "21a4b071-b7f0-4b74-8ccb-740eef31f462" (UID: "21a4b071-b7f0-4b74-8ccb-740eef31f462"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.179721 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a4b071-b7f0-4b74-8ccb-740eef31f462-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.182925 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a4b071-b7f0-4b74-8ccb-740eef31f462-kube-api-access-2bg7h" (OuterVolumeSpecName: "kube-api-access-2bg7h") pod "21a4b071-b7f0-4b74-8ccb-740eef31f462" (UID: "21a4b071-b7f0-4b74-8ccb-740eef31f462"). InnerVolumeSpecName "kube-api-access-2bg7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.207478 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-config-data" (OuterVolumeSpecName: "config-data") pod "21a4b071-b7f0-4b74-8ccb-740eef31f462" (UID: "21a4b071-b7f0-4b74-8ccb-740eef31f462"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.228388 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21a4b071-b7f0-4b74-8ccb-740eef31f462" (UID: "21a4b071-b7f0-4b74-8ccb-740eef31f462"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.282184 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.282510 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a4b071-b7f0-4b74-8ccb-740eef31f462-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.282679 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bg7h\" (UniqueName: \"kubernetes.io/projected/21a4b071-b7f0-4b74-8ccb-740eef31f462-kube-api-access-2bg7h\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.934971 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.935009 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21a4b071-b7f0-4b74-8ccb-740eef31f462","Type":"ContainerDied","Data":"8b437154174b16d8abc631d94aecd568c27b1810196feb017ae94f979af754e3"} Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.936443 4796 scope.go:117] "RemoveContainer" containerID="4ba7d8e46301d427593616a3116554fb5ee8715265eee58d02e4204da21ec87a" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.947301 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerStarted","Data":"1a3dab3268132c4f7eea1f8bcedff7f8cc2481a1e7efb83a060dc17002a040c9"} Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.963447 4796 scope.go:117] "RemoveContainer" containerID="5b27df5a83b802e33c8a4c14826b32e167bae1d47891b1f274714eed193ea090" Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.986881 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:40 crc kubenswrapper[4796]: I1127 11:58:40.996820 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.006045 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:41 crc kubenswrapper[4796]: E1127 11:58:41.006449 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-log" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.006466 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-log" Nov 27 11:58:41 crc kubenswrapper[4796]: E1127 11:58:41.006485 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-metadata" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.006492 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-metadata" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.006678 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-metadata" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.006693 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" containerName="nova-metadata-log" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.007641 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.015671 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.015891 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.024950 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.201996 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-config-data\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.202427 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.202559 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.202598 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f129cfd3-1997-42cd-aea7-f99e796c2351-logs\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.202638 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmt6f\" (UniqueName: \"kubernetes.io/projected/f129cfd3-1997-42cd-aea7-f99e796c2351-kube-api-access-xmt6f\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.304337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.304608 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f129cfd3-1997-42cd-aea7-f99e796c2351-logs\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.304754 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmt6f\" (UniqueName: \"kubernetes.io/projected/f129cfd3-1997-42cd-aea7-f99e796c2351-kube-api-access-xmt6f\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.304853 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-config-data\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.304992 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.305067 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f129cfd3-1997-42cd-aea7-f99e796c2351-logs\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.313033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-config-data\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.317926 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.319982 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.335953 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmt6f\" (UniqueName: \"kubernetes.io/projected/f129cfd3-1997-42cd-aea7-f99e796c2351-kube-api-access-xmt6f\") pod \"nova-metadata-0\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.583261 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a4b071-b7f0-4b74-8ccb-740eef31f462" path="/var/lib/kubelet/pods/21a4b071-b7f0-4b74-8ccb-740eef31f462/volumes" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.630032 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.957704 4796 generic.go:334] "Generic (PLEG): container finished" podID="9cbdc050-92dd-44be-83f2-474365084f53" containerID="eac3bfb17e80603b35583d733348e26451d16279d1d6bd997f982c96b2f362a6" exitCode=0 Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.957804 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lbg4j" event={"ID":"9cbdc050-92dd-44be-83f2-474365084f53","Type":"ContainerDied","Data":"eac3bfb17e80603b35583d733348e26451d16279d1d6bd997f982c96b2f362a6"} Nov 27 11:58:41 crc kubenswrapper[4796]: I1127 11:58:41.962095 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerStarted","Data":"661989fb593bf91f4379032d8a1ad87bed6d17f13f04753aefd24449606f5b04"} Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.113530 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:42 crc kubenswrapper[4796]: W1127 11:58:42.115261 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf129cfd3_1997_42cd_aea7_f99e796c2351.slice/crio-3d40022091746c3142855e690a006462d48d504c5d7c35604ff49c0007106849 WatchSource:0}: Error finding container 3d40022091746c3142855e690a006462d48d504c5d7c35604ff49c0007106849: Status 404 returned error can't find the container with id 3d40022091746c3142855e690a006462d48d504c5d7c35604ff49c0007106849 Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.318861 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.318909 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.550250 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.550642 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.585773 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.604601 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.729568 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.815328 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-pkmhj"] Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.815596 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerName="dnsmasq-dns" containerID="cri-o://049434f9a8690be59e451762d02797a69d0a8251bd0e920c60965444600a5c0b" gracePeriod=10 Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.972143 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f129cfd3-1997-42cd-aea7-f99e796c2351","Type":"ContainerStarted","Data":"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b"} Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.972183 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f129cfd3-1997-42cd-aea7-f99e796c2351","Type":"ContainerStarted","Data":"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0"} Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.972191 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f129cfd3-1997-42cd-aea7-f99e796c2351","Type":"ContainerStarted","Data":"3d40022091746c3142855e690a006462d48d504c5d7c35604ff49c0007106849"} Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.974961 4796 generic.go:334] "Generic (PLEG): container finished" podID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerID="049434f9a8690be59e451762d02797a69d0a8251bd0e920c60965444600a5c0b" exitCode=0 Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.975002 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" event={"ID":"8cdbd3cc-dbee-4abb-8e34-c910554b5142","Type":"ContainerDied","Data":"049434f9a8690be59e451762d02797a69d0a8251bd0e920c60965444600a5c0b"} Nov 27 11:58:42 crc kubenswrapper[4796]: I1127 11:58:42.996810 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.996792358 podStartE2EDuration="2.996792358s" podCreationTimestamp="2025-11-27 11:58:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:42.988613345 +0000 UTC m=+2040.506932263" watchObservedRunningTime="2025-11-27 11:58:42.996792358 +0000 UTC m=+2040.515111276" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.017665 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.404664 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.405145 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.495401 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.511647 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555158 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jjhj\" (UniqueName: \"kubernetes.io/projected/9cbdc050-92dd-44be-83f2-474365084f53-kube-api-access-4jjhj\") pod \"9cbdc050-92dd-44be-83f2-474365084f53\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555219 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data\") pod \"9cbdc050-92dd-44be-83f2-474365084f53\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555357 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-svc\") pod \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555397 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-scripts\") pod \"9cbdc050-92dd-44be-83f2-474365084f53\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555454 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dt4n\" (UniqueName: \"kubernetes.io/projected/8cdbd3cc-dbee-4abb-8e34-c910554b5142-kube-api-access-9dt4n\") pod \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555529 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-sb\") pod \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555572 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-combined-ca-bundle\") pod \"9cbdc050-92dd-44be-83f2-474365084f53\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555608 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-config\") pod \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555651 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-swift-storage-0\") pod \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.555698 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-nb\") pod \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\" (UID: \"8cdbd3cc-dbee-4abb-8e34-c910554b5142\") " Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.579920 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cbdc050-92dd-44be-83f2-474365084f53-kube-api-access-4jjhj" (OuterVolumeSpecName: "kube-api-access-4jjhj") pod "9cbdc050-92dd-44be-83f2-474365084f53" (UID: "9cbdc050-92dd-44be-83f2-474365084f53"). InnerVolumeSpecName "kube-api-access-4jjhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.600428 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-scripts" (OuterVolumeSpecName: "scripts") pod "9cbdc050-92dd-44be-83f2-474365084f53" (UID: "9cbdc050-92dd-44be-83f2-474365084f53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.606547 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cdbd3cc-dbee-4abb-8e34-c910554b5142-kube-api-access-9dt4n" (OuterVolumeSpecName: "kube-api-access-9dt4n") pod "8cdbd3cc-dbee-4abb-8e34-c910554b5142" (UID: "8cdbd3cc-dbee-4abb-8e34-c910554b5142"). InnerVolumeSpecName "kube-api-access-9dt4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.652941 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-config" (OuterVolumeSpecName: "config") pod "8cdbd3cc-dbee-4abb-8e34-c910554b5142" (UID: "8cdbd3cc-dbee-4abb-8e34-c910554b5142"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.658448 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data" (OuterVolumeSpecName: "config-data") pod "9cbdc050-92dd-44be-83f2-474365084f53" (UID: "9cbdc050-92dd-44be-83f2-474365084f53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.659034 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data\") pod \"9cbdc050-92dd-44be-83f2-474365084f53\" (UID: \"9cbdc050-92dd-44be-83f2-474365084f53\") " Nov 27 11:58:43 crc kubenswrapper[4796]: W1127 11:58:43.659516 4796 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9cbdc050-92dd-44be-83f2-474365084f53/volumes/kubernetes.io~secret/config-data Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.659551 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data" (OuterVolumeSpecName: "config-data") pod "9cbdc050-92dd-44be-83f2-474365084f53" (UID: "9cbdc050-92dd-44be-83f2-474365084f53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.661467 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jjhj\" (UniqueName: \"kubernetes.io/projected/9cbdc050-92dd-44be-83f2-474365084f53-kube-api-access-4jjhj\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.661488 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.661497 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.661506 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dt4n\" (UniqueName: \"kubernetes.io/projected/8cdbd3cc-dbee-4abb-8e34-c910554b5142-kube-api-access-9dt4n\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.661516 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.692566 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cbdc050-92dd-44be-83f2-474365084f53" (UID: "9cbdc050-92dd-44be-83f2-474365084f53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.725696 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8cdbd3cc-dbee-4abb-8e34-c910554b5142" (UID: "8cdbd3cc-dbee-4abb-8e34-c910554b5142"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.725810 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8cdbd3cc-dbee-4abb-8e34-c910554b5142" (UID: "8cdbd3cc-dbee-4abb-8e34-c910554b5142"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.739371 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8cdbd3cc-dbee-4abb-8e34-c910554b5142" (UID: "8cdbd3cc-dbee-4abb-8e34-c910554b5142"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.745888 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8cdbd3cc-dbee-4abb-8e34-c910554b5142" (UID: "8cdbd3cc-dbee-4abb-8e34-c910554b5142"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.763531 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cbdc050-92dd-44be-83f2-474365084f53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.763612 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.763627 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.763640 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.763659 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cdbd3cc-dbee-4abb-8e34-c910554b5142-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.992162 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" event={"ID":"8cdbd3cc-dbee-4abb-8e34-c910554b5142","Type":"ContainerDied","Data":"99bdb254076ef589d2b316193793ce17d4368950acd201c2b5330a039ed9b306"} Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.992215 4796 scope.go:117] "RemoveContainer" containerID="049434f9a8690be59e451762d02797a69d0a8251bd0e920c60965444600a5c0b" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.992364 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-pkmhj" Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.999577 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerStarted","Data":"f58112ad4949b195b7e25400bb8a0a3889a8420c80260c5293861185fe26ac3d"} Nov 27 11:58:43 crc kubenswrapper[4796]: I1127 11:58:43.999758 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.007357 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lbg4j" event={"ID":"9cbdc050-92dd-44be-83f2-474365084f53","Type":"ContainerDied","Data":"45a34489e8ea5cb39113a2bc12c0762842258297a4b558766d022cbbafbf07c2"} Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.007605 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45a34489e8ea5cb39113a2bc12c0762842258297a4b558766d022cbbafbf07c2" Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.007778 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lbg4j" Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.020151 4796 scope.go:117] "RemoveContainer" containerID="fef71d3d94aab373167ea9ac4c5d3279393a5dd18601243f21ec3aa4741d5f29" Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.057279 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.394520367 podStartE2EDuration="6.057245534s" podCreationTimestamp="2025-11-27 11:58:38 +0000 UTC" firstStartedPulling="2025-11-27 11:58:39.034367962 +0000 UTC m=+2036.552686890" lastFinishedPulling="2025-11-27 11:58:42.697093149 +0000 UTC m=+2040.215412057" observedRunningTime="2025-11-27 11:58:44.038198605 +0000 UTC m=+2041.556517523" watchObservedRunningTime="2025-11-27 11:58:44.057245534 +0000 UTC m=+2041.575564452" Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.084121 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-pkmhj"] Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.093522 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-pkmhj"] Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.169170 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.169410 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-log" containerID="cri-o://467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7" gracePeriod=30 Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.169859 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-api" containerID="cri-o://781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f" gracePeriod=30 Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.202117 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:44 crc kubenswrapper[4796]: I1127 11:58:44.213878 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:44 crc kubenswrapper[4796]: E1127 11:58:44.307068 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa3bdadd_98ed_4f1e_88c2_d23354a0e84a.slice/crio-5593861da04285f8c0fd914fe3b491e2e50de627ed3718f78d306a76e8e08b28.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cf8de36_eb5f_4cc1_b8ba_820ded14bb85.slice/crio-conmon-467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa3bdadd_98ed_4f1e_88c2_d23354a0e84a.slice/crio-conmon-5593861da04285f8c0fd914fe3b491e2e50de627ed3718f78d306a76e8e08b28.scope\": RecentStats: unable to find data in memory cache]" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.020727 4796 generic.go:334] "Generic (PLEG): container finished" podID="aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" containerID="5593861da04285f8c0fd914fe3b491e2e50de627ed3718f78d306a76e8e08b28" exitCode=0 Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.020769 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" event={"ID":"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a","Type":"ContainerDied","Data":"5593861da04285f8c0fd914fe3b491e2e50de627ed3718f78d306a76e8e08b28"} Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.024759 4796 generic.go:334] "Generic (PLEG): container finished" podID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerID="467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7" exitCode=143 Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.025017 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="69926936-b4f5-4201-8df2-3ac92004ad81" containerName="nova-scheduler-scheduler" containerID="cri-o://db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519" gracePeriod=30 Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.025548 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85","Type":"ContainerDied","Data":"467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7"} Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.027192 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-log" containerID="cri-o://a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0" gracePeriod=30 Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.027402 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-metadata" containerID="cri-o://e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b" gracePeriod=30 Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.569712 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.581963 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" path="/var/lib/kubelet/pods/8cdbd3cc-dbee-4abb-8e34-c910554b5142/volumes" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.798876 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.807990 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-combined-ca-bundle\") pod \"f129cfd3-1997-42cd-aea7-f99e796c2351\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.808036 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-nova-metadata-tls-certs\") pod \"f129cfd3-1997-42cd-aea7-f99e796c2351\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.808065 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmt6f\" (UniqueName: \"kubernetes.io/projected/f129cfd3-1997-42cd-aea7-f99e796c2351-kube-api-access-xmt6f\") pod \"f129cfd3-1997-42cd-aea7-f99e796c2351\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.808085 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-config-data\") pod \"f129cfd3-1997-42cd-aea7-f99e796c2351\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.808143 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f129cfd3-1997-42cd-aea7-f99e796c2351-logs\") pod \"f129cfd3-1997-42cd-aea7-f99e796c2351\" (UID: \"f129cfd3-1997-42cd-aea7-f99e796c2351\") " Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.808767 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f129cfd3-1997-42cd-aea7-f99e796c2351-logs" (OuterVolumeSpecName: "logs") pod "f129cfd3-1997-42cd-aea7-f99e796c2351" (UID: "f129cfd3-1997-42cd-aea7-f99e796c2351"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.847736 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f129cfd3-1997-42cd-aea7-f99e796c2351-kube-api-access-xmt6f" (OuterVolumeSpecName: "kube-api-access-xmt6f") pod "f129cfd3-1997-42cd-aea7-f99e796c2351" (UID: "f129cfd3-1997-42cd-aea7-f99e796c2351"). InnerVolumeSpecName "kube-api-access-xmt6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.889181 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f129cfd3-1997-42cd-aea7-f99e796c2351" (UID: "f129cfd3-1997-42cd-aea7-f99e796c2351"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.911344 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-config-data" (OuterVolumeSpecName: "config-data") pod "f129cfd3-1997-42cd-aea7-f99e796c2351" (UID: "f129cfd3-1997-42cd-aea7-f99e796c2351"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.911756 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.911774 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmt6f\" (UniqueName: \"kubernetes.io/projected/f129cfd3-1997-42cd-aea7-f99e796c2351-kube-api-access-xmt6f\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.911786 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.911795 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f129cfd3-1997-42cd-aea7-f99e796c2351-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.931850 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f129cfd3-1997-42cd-aea7-f99e796c2351" (UID: "f129cfd3-1997-42cd-aea7-f99e796c2351"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:45 crc kubenswrapper[4796]: I1127 11:58:45.956352 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.013235 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-config-data\") pod \"69926936-b4f5-4201-8df2-3ac92004ad81\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.013341 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fqdt\" (UniqueName: \"kubernetes.io/projected/69926936-b4f5-4201-8df2-3ac92004ad81-kube-api-access-4fqdt\") pod \"69926936-b4f5-4201-8df2-3ac92004ad81\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.013404 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-combined-ca-bundle\") pod \"69926936-b4f5-4201-8df2-3ac92004ad81\" (UID: \"69926936-b4f5-4201-8df2-3ac92004ad81\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.013793 4796 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f129cfd3-1997-42cd-aea7-f99e796c2351-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.020434 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69926936-b4f5-4201-8df2-3ac92004ad81-kube-api-access-4fqdt" (OuterVolumeSpecName: "kube-api-access-4fqdt") pod "69926936-b4f5-4201-8df2-3ac92004ad81" (UID: "69926936-b4f5-4201-8df2-3ac92004ad81"). InnerVolumeSpecName "kube-api-access-4fqdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038095 4796 generic.go:334] "Generic (PLEG): container finished" podID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerID="e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b" exitCode=0 Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038130 4796 generic.go:334] "Generic (PLEG): container finished" podID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerID="a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0" exitCode=143 Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038203 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f129cfd3-1997-42cd-aea7-f99e796c2351","Type":"ContainerDied","Data":"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b"} Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038197 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038233 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f129cfd3-1997-42cd-aea7-f99e796c2351","Type":"ContainerDied","Data":"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0"} Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038248 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f129cfd3-1997-42cd-aea7-f99e796c2351","Type":"ContainerDied","Data":"3d40022091746c3142855e690a006462d48d504c5d7c35604ff49c0007106849"} Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.038285 4796 scope.go:117] "RemoveContainer" containerID="e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.046192 4796 generic.go:334] "Generic (PLEG): container finished" podID="69926936-b4f5-4201-8df2-3ac92004ad81" containerID="db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519" exitCode=0 Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.048775 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69926936-b4f5-4201-8df2-3ac92004ad81","Type":"ContainerDied","Data":"db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519"} Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.049737 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69926936-b4f5-4201-8df2-3ac92004ad81","Type":"ContainerDied","Data":"e9b21e9bd04e9994f2d83421b0d2dcd017afb17dcf8bc152225ce11106513f8d"} Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.048839 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.056344 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"8d02aea19dfdaf4e45b9254b7fcf0891a26f4b7262a59d6766b8fb31a94cc4ee"} Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.059840 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-config-data" (OuterVolumeSpecName: "config-data") pod "69926936-b4f5-4201-8df2-3ac92004ad81" (UID: "69926936-b4f5-4201-8df2-3ac92004ad81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.064584 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69926936-b4f5-4201-8df2-3ac92004ad81" (UID: "69926936-b4f5-4201-8df2-3ac92004ad81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.087476 4796 scope.go:117] "RemoveContainer" containerID="a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.098686 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.118282 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.118311 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69926936-b4f5-4201-8df2-3ac92004ad81-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.118322 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fqdt\" (UniqueName: \"kubernetes.io/projected/69926936-b4f5-4201-8df2-3ac92004ad81-kube-api-access-4fqdt\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.137628 4796 scope.go:117] "RemoveContainer" containerID="e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.137762 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.140561 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b\": container with ID starting with e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b not found: ID does not exist" containerID="e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.140610 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b"} err="failed to get container status \"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b\": rpc error: code = NotFound desc = could not find container \"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b\": container with ID starting with e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b not found: ID does not exist" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.140639 4796 scope.go:117] "RemoveContainer" containerID="a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.141037 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0\": container with ID starting with a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0 not found: ID does not exist" containerID="a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.141084 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0"} err="failed to get container status \"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0\": rpc error: code = NotFound desc = could not find container \"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0\": container with ID starting with a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0 not found: ID does not exist" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.141113 4796 scope.go:117] "RemoveContainer" containerID="e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.141843 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b"} err="failed to get container status \"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b\": rpc error: code = NotFound desc = could not find container \"e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b\": container with ID starting with e602eed1a727d5fab69eb9566266aa39d17e4b064a8329dfefbc660333f45e9b not found: ID does not exist" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.141871 4796 scope.go:117] "RemoveContainer" containerID="a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.142203 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0"} err="failed to get container status \"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0\": rpc error: code = NotFound desc = could not find container \"a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0\": container with ID starting with a1e37e9863e3b698879751ca5788d4350e3c129491ee2672cb3efe4e1e147dc0 not found: ID does not exist" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.142223 4796 scope.go:117] "RemoveContainer" containerID="db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.177445 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.194859 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69926936-b4f5-4201-8df2-3ac92004ad81" containerName="nova-scheduler-scheduler" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.194901 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="69926936-b4f5-4201-8df2-3ac92004ad81" containerName="nova-scheduler-scheduler" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.194929 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerName="init" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.194936 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerName="init" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.194964 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-metadata" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.194970 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-metadata" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.194982 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerName="dnsmasq-dns" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.194988 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerName="dnsmasq-dns" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.195030 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbdc050-92dd-44be-83f2-474365084f53" containerName="nova-manage" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195037 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbdc050-92dd-44be-83f2-474365084f53" containerName="nova-manage" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.195047 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-log" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195054 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-log" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195399 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-metadata" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195415 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cdbd3cc-dbee-4abb-8e34-c910554b5142" containerName="dnsmasq-dns" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195424 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbdc050-92dd-44be-83f2-474365084f53" containerName="nova-manage" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195436 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="69926936-b4f5-4201-8df2-3ac92004ad81" containerName="nova-scheduler-scheduler" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.195447 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" containerName="nova-metadata-log" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.196385 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.196467 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.197866 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.198640 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.217867 4796 scope.go:117] "RemoveContainer" containerID="db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519" Nov 27 11:58:46 crc kubenswrapper[4796]: E1127 11:58:46.218336 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519\": container with ID starting with db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519 not found: ID does not exist" containerID="db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.218378 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519"} err="failed to get container status \"db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519\": rpc error: code = NotFound desc = could not find container \"db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519\": container with ID starting with db4af2ab4a83b65a4c0996195a568eea73f74eb99c068282f3cf5491f3733519 not found: ID does not exist" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.225783 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-config-data\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.225923 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.225997 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5c6984-1064-4944-a335-9cd8fa5c1738-logs\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.226076 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx4rv\" (UniqueName: \"kubernetes.io/projected/5a5c6984-1064-4944-a335-9cd8fa5c1738-kube-api-access-gx4rv\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.226098 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.327829 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-config-data\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.327910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.327947 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5c6984-1064-4944-a335-9cd8fa5c1738-logs\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.327986 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx4rv\" (UniqueName: \"kubernetes.io/projected/5a5c6984-1064-4944-a335-9cd8fa5c1738-kube-api-access-gx4rv\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.328005 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.328812 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5c6984-1064-4944-a335-9cd8fa5c1738-logs\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.331108 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-config-data\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.334560 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.338718 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.347547 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx4rv\" (UniqueName: \"kubernetes.io/projected/5a5c6984-1064-4944-a335-9cd8fa5c1738-kube-api-access-gx4rv\") pod \"nova-metadata-0\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.388018 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.401082 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.412987 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.414172 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.424172 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.424715 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.429300 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-config-data\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.429370 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvbh9\" (UniqueName: \"kubernetes.io/projected/406db992-3094-4324-a2fb-ed288d3ee288-kube-api-access-lvbh9\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.429426 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.473359 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.520178 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.531131 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-config-data\") pod \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.531223 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwmlq\" (UniqueName: \"kubernetes.io/projected/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-kube-api-access-jwmlq\") pod \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.531346 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-combined-ca-bundle\") pod \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.531426 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-scripts\") pod \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\" (UID: \"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a\") " Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.531786 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvbh9\" (UniqueName: \"kubernetes.io/projected/406db992-3094-4324-a2fb-ed288d3ee288-kube-api-access-lvbh9\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.531849 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.532009 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-config-data\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.536603 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-kube-api-access-jwmlq" (OuterVolumeSpecName: "kube-api-access-jwmlq") pod "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" (UID: "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a"). InnerVolumeSpecName "kube-api-access-jwmlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.536848 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-config-data\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.538201 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-scripts" (OuterVolumeSpecName: "scripts") pod "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" (UID: "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.544804 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.559252 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvbh9\" (UniqueName: \"kubernetes.io/projected/406db992-3094-4324-a2fb-ed288d3ee288-kube-api-access-lvbh9\") pod \"nova-scheduler-0\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " pod="openstack/nova-scheduler-0" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.566128 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" (UID: "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.582490 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-config-data" (OuterVolumeSpecName: "config-data") pod "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" (UID: "aa3bdadd-98ed-4f1e-88c2-d23354a0e84a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.633816 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.634103 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.634115 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwmlq\" (UniqueName: \"kubernetes.io/projected/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-kube-api-access-jwmlq\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.634125 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:46 crc kubenswrapper[4796]: I1127 11:58:46.741084 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.014443 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.088077 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5a5c6984-1064-4944-a335-9cd8fa5c1738","Type":"ContainerStarted","Data":"588d6740ab97c59d37dba741e099bef9274db88c3b3a62ad5cdd2f6f19c225c9"} Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.089774 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" event={"ID":"aa3bdadd-98ed-4f1e-88c2-d23354a0e84a","Type":"ContainerDied","Data":"c0f6fe650d76ea1283fe0b6ecc25959b4f85bb258315cc7fc3eabdb68fe55388"} Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.089810 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f6fe650d76ea1283fe0b6ecc25959b4f85bb258315cc7fc3eabdb68fe55388" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.089856 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nzgfc" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.141049 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 11:58:47 crc kubenswrapper[4796]: E1127 11:58:47.141474 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" containerName="nova-cell1-conductor-db-sync" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.141490 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" containerName="nova-cell1-conductor-db-sync" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.141924 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" containerName="nova-cell1-conductor-db-sync" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.143298 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.149015 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.165336 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.203503 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.245145 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d375ad-8df3-4053-a31a-1483cbc5ce30-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.245316 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d375ad-8df3-4053-a31a-1483cbc5ce30-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.245493 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/97d375ad-8df3-4053-a31a-1483cbc5ce30-kube-api-access-xqkk5\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.347276 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d375ad-8df3-4053-a31a-1483cbc5ce30-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.347352 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/97d375ad-8df3-4053-a31a-1483cbc5ce30-kube-api-access-xqkk5\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.347381 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d375ad-8df3-4053-a31a-1483cbc5ce30-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.356128 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d375ad-8df3-4053-a31a-1483cbc5ce30-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.362902 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d375ad-8df3-4053-a31a-1483cbc5ce30-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.384818 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/97d375ad-8df3-4053-a31a-1483cbc5ce30-kube-api-access-xqkk5\") pod \"nova-cell1-conductor-0\" (UID: \"97d375ad-8df3-4053-a31a-1483cbc5ce30\") " pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.460759 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.615504 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69926936-b4f5-4201-8df2-3ac92004ad81" path="/var/lib/kubelet/pods/69926936-b4f5-4201-8df2-3ac92004ad81/volumes" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.618639 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f129cfd3-1997-42cd-aea7-f99e796c2351" path="/var/lib/kubelet/pods/f129cfd3-1997-42cd-aea7-f99e796c2351/volumes" Nov 27 11:58:47 crc kubenswrapper[4796]: I1127 11:58:47.937743 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 11:58:47 crc kubenswrapper[4796]: W1127 11:58:47.942566 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97d375ad_8df3_4053_a31a_1483cbc5ce30.slice/crio-95f6aeff7c20b4f7ba0ec1e8a3635cc1487cd199e9402f812798228fdf949420 WatchSource:0}: Error finding container 95f6aeff7c20b4f7ba0ec1e8a3635cc1487cd199e9402f812798228fdf949420: Status 404 returned error can't find the container with id 95f6aeff7c20b4f7ba0ec1e8a3635cc1487cd199e9402f812798228fdf949420 Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.121977 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406db992-3094-4324-a2fb-ed288d3ee288","Type":"ContainerStarted","Data":"df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d"} Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.122089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406db992-3094-4324-a2fb-ed288d3ee288","Type":"ContainerStarted","Data":"ff741c9701221d4513d8a19142261568a74481c6a1629220fa949352bd6203c6"} Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.131062 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5a5c6984-1064-4944-a335-9cd8fa5c1738","Type":"ContainerStarted","Data":"83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026"} Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.131107 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5a5c6984-1064-4944-a335-9cd8fa5c1738","Type":"ContainerStarted","Data":"796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254"} Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.135259 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"97d375ad-8df3-4053-a31a-1483cbc5ce30","Type":"ContainerStarted","Data":"95f6aeff7c20b4f7ba0ec1e8a3635cc1487cd199e9402f812798228fdf949420"} Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.144788 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.14477066 podStartE2EDuration="2.14477066s" podCreationTimestamp="2025-11-27 11:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:48.141485761 +0000 UTC m=+2045.659804679" watchObservedRunningTime="2025-11-27 11:58:48.14477066 +0000 UTC m=+2045.663089578" Nov 27 11:58:48 crc kubenswrapper[4796]: I1127 11:58:48.179988 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.179966409 podStartE2EDuration="2.179966409s" podCreationTimestamp="2025-11-27 11:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:48.159588774 +0000 UTC m=+2045.677907692" watchObservedRunningTime="2025-11-27 11:58:48.179966409 +0000 UTC m=+2045.698285327" Nov 27 11:58:49 crc kubenswrapper[4796]: I1127 11:58:49.155335 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"97d375ad-8df3-4053-a31a-1483cbc5ce30","Type":"ContainerStarted","Data":"db6a6164f9a785a4670f8495e3dc6c79039bffe4d962759ac7c1aff7827ff230"} Nov 27 11:58:49 crc kubenswrapper[4796]: I1127 11:58:49.191856 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.191828121 podStartE2EDuration="2.191828121s" podCreationTimestamp="2025-11-27 11:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:49.17676284 +0000 UTC m=+2046.695081788" watchObservedRunningTime="2025-11-27 11:58:49.191828121 +0000 UTC m=+2046.710147049" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.142454 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.166497 4796 generic.go:334] "Generic (PLEG): container finished" podID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerID="781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f" exitCode=0 Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.166587 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.166605 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85","Type":"ContainerDied","Data":"781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f"} Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.167615 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85","Type":"ContainerDied","Data":"fe5ffeecb664613185ab88f85f6ebd77a72be342a8235ea5ba6f6fd4bdebfdff"} Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.167647 4796 scope.go:117] "RemoveContainer" containerID="781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.168135 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.203108 4796 scope.go:117] "RemoveContainer" containerID="467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.222848 4796 scope.go:117] "RemoveContainer" containerID="781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f" Nov 27 11:58:50 crc kubenswrapper[4796]: E1127 11:58:50.223181 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f\": container with ID starting with 781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f not found: ID does not exist" containerID="781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.223213 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f"} err="failed to get container status \"781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f\": rpc error: code = NotFound desc = could not find container \"781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f\": container with ID starting with 781f9be52f62ad3d0e607279400113b5651602ffa290d03b293d0be3f3c0495f not found: ID does not exist" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.223231 4796 scope.go:117] "RemoveContainer" containerID="467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7" Nov 27 11:58:50 crc kubenswrapper[4796]: E1127 11:58:50.223590 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7\": container with ID starting with 467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7 not found: ID does not exist" containerID="467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.223704 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7"} err="failed to get container status \"467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7\": rpc error: code = NotFound desc = could not find container \"467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7\": container with ID starting with 467a629aae3a910358a9bbc2a958ec37938d26db5c664eb1071d60db092292d7 not found: ID does not exist" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.304026 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data\") pod \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.304098 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg8tg\" (UniqueName: \"kubernetes.io/projected/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-kube-api-access-xg8tg\") pod \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.304157 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-logs\") pod \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.304292 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-combined-ca-bundle\") pod \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.304828 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-logs" (OuterVolumeSpecName: "logs") pod "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" (UID: "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.321639 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-kube-api-access-xg8tg" (OuterVolumeSpecName: "kube-api-access-xg8tg") pod "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" (UID: "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85"). InnerVolumeSpecName "kube-api-access-xg8tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:58:50 crc kubenswrapper[4796]: E1127 11:58:50.358570 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data podName:3cf8de36-eb5f-4cc1-b8ba-820ded14bb85 nodeName:}" failed. No retries permitted until 2025-11-27 11:58:50.858539992 +0000 UTC m=+2048.376858920 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data") pod "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" (UID: "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85") : error deleting /var/lib/kubelet/pods/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85/volume-subpaths: remove /var/lib/kubelet/pods/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85/volume-subpaths: no such file or directory Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.362940 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" (UID: "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.406904 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.406940 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg8tg\" (UniqueName: \"kubernetes.io/projected/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-kube-api-access-xg8tg\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.406955 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.915964 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data\") pod \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\" (UID: \"3cf8de36-eb5f-4cc1-b8ba-820ded14bb85\") " Nov 27 11:58:50 crc kubenswrapper[4796]: I1127 11:58:50.920448 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data" (OuterVolumeSpecName: "config-data") pod "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" (UID: "3cf8de36-eb5f-4cc1-b8ba-820ded14bb85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.018225 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.116304 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.147961 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.165511 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:51 crc kubenswrapper[4796]: E1127 11:58:51.166006 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-log" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.166022 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-log" Nov 27 11:58:51 crc kubenswrapper[4796]: E1127 11:58:51.166036 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-api" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.166041 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-api" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.166210 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-log" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.166235 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" containerName="nova-api-api" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.170318 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.173337 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.188008 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.325171 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cljp2\" (UniqueName: \"kubernetes.io/projected/3017636c-aebd-45bf-a21c-5cee6583369d-kube-api-access-cljp2\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.325228 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3017636c-aebd-45bf-a21c-5cee6583369d-logs\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.325387 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.325734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-config-data\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.428036 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.428426 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-config-data\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.429439 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cljp2\" (UniqueName: \"kubernetes.io/projected/3017636c-aebd-45bf-a21c-5cee6583369d-kube-api-access-cljp2\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.429518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3017636c-aebd-45bf-a21c-5cee6583369d-logs\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.430210 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3017636c-aebd-45bf-a21c-5cee6583369d-logs\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.434546 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-config-data\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.436441 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.458572 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cljp2\" (UniqueName: \"kubernetes.io/projected/3017636c-aebd-45bf-a21c-5cee6583369d-kube-api-access-cljp2\") pod \"nova-api-0\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.498351 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.521351 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.521437 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.589715 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf8de36-eb5f-4cc1-b8ba-820ded14bb85" path="/var/lib/kubelet/pods/3cf8de36-eb5f-4cc1-b8ba-820ded14bb85/volumes" Nov 27 11:58:51 crc kubenswrapper[4796]: I1127 11:58:51.741803 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 11:58:52 crc kubenswrapper[4796]: I1127 11:58:52.006690 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:58:52 crc kubenswrapper[4796]: I1127 11:58:52.226868 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3017636c-aebd-45bf-a21c-5cee6583369d","Type":"ContainerStarted","Data":"82711683561e67c4ef1cbd66ebf23b9f161f6a47d40df81efced6fc104a13cde"} Nov 27 11:58:52 crc kubenswrapper[4796]: I1127 11:58:52.227224 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3017636c-aebd-45bf-a21c-5cee6583369d","Type":"ContainerStarted","Data":"e5fa804ed46c13507e4aac78a587d093b4aca82698e97ef9b12ed461e0d6ac4b"} Nov 27 11:58:53 crc kubenswrapper[4796]: I1127 11:58:53.241027 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3017636c-aebd-45bf-a21c-5cee6583369d","Type":"ContainerStarted","Data":"ae0f76141761dc767260ee17ed69999da02ae0f432b6ee873c35f645bcfb025b"} Nov 27 11:58:53 crc kubenswrapper[4796]: I1127 11:58:53.270236 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.270215538 podStartE2EDuration="2.270215538s" podCreationTimestamp="2025-11-27 11:58:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:58:53.259048363 +0000 UTC m=+2050.777367281" watchObservedRunningTime="2025-11-27 11:58:53.270215538 +0000 UTC m=+2050.788534466" Nov 27 11:58:56 crc kubenswrapper[4796]: I1127 11:58:56.520703 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 11:58:56 crc kubenswrapper[4796]: I1127 11:58:56.521023 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 11:58:56 crc kubenswrapper[4796]: I1127 11:58:56.741996 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 11:58:56 crc kubenswrapper[4796]: I1127 11:58:56.769599 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 11:58:57 crc kubenswrapper[4796]: I1127 11:58:57.334218 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 11:58:57 crc kubenswrapper[4796]: I1127 11:58:57.519369 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 27 11:58:57 crc kubenswrapper[4796]: I1127 11:58:57.530428 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 11:58:57 crc kubenswrapper[4796]: I1127 11:58:57.530441 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:01 crc kubenswrapper[4796]: I1127 11:59:01.498748 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 11:59:01 crc kubenswrapper[4796]: I1127 11:59:01.499292 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 11:59:02 crc kubenswrapper[4796]: I1127 11:59:02.581519 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:02 crc kubenswrapper[4796]: I1127 11:59:02.581551 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:06 crc kubenswrapper[4796]: I1127 11:59:06.526399 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 11:59:06 crc kubenswrapper[4796]: I1127 11:59:06.528429 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 11:59:06 crc kubenswrapper[4796]: I1127 11:59:06.532975 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 11:59:07 crc kubenswrapper[4796]: I1127 11:59:07.434693 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 11:59:08 crc kubenswrapper[4796]: I1127 11:59:08.441052 4796 generic.go:334] "Generic (PLEG): container finished" podID="22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" containerID="46170f34192949854fbd2254a59b79816457c154546193bcc9a421c252b70e9f" exitCode=137 Nov 27 11:59:08 crc kubenswrapper[4796]: I1127 11:59:08.441152 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c","Type":"ContainerDied","Data":"46170f34192949854fbd2254a59b79816457c154546193bcc9a421c252b70e9f"} Nov 27 11:59:08 crc kubenswrapper[4796]: I1127 11:59:08.555936 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.008257 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.123042 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-combined-ca-bundle\") pod \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.123409 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-config-data\") pod \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.123555 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrrx7\" (UniqueName: \"kubernetes.io/projected/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-kube-api-access-vrrx7\") pod \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\" (UID: \"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c\") " Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.129976 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-kube-api-access-vrrx7" (OuterVolumeSpecName: "kube-api-access-vrrx7") pod "22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" (UID: "22ba5d61-38de-4bae-ab5e-3922e3c7fb9c"). InnerVolumeSpecName "kube-api-access-vrrx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.162885 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" (UID: "22ba5d61-38de-4bae-ab5e-3922e3c7fb9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.163402 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-config-data" (OuterVolumeSpecName: "config-data") pod "22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" (UID: "22ba5d61-38de-4bae-ab5e-3922e3c7fb9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.225404 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.225440 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.225471 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrrx7\" (UniqueName: \"kubernetes.io/projected/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c-kube-api-access-vrrx7\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.455367 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"22ba5d61-38de-4bae-ab5e-3922e3c7fb9c","Type":"ContainerDied","Data":"f8d514fbff7adc2b770468c973bc81374e33a05fc6df86f0cd2084858620afc9"} Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.455400 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.455420 4796 scope.go:117] "RemoveContainer" containerID="46170f34192949854fbd2254a59b79816457c154546193bcc9a421c252b70e9f" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.488642 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.501625 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.527911 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:59:09 crc kubenswrapper[4796]: E1127 11:59:09.528323 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.528338 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.528504 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.529105 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.536650 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.536811 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.537088 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.540008 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.579998 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22ba5d61-38de-4bae-ab5e-3922e3c7fb9c" path="/var/lib/kubelet/pods/22ba5d61-38de-4bae-ab5e-3922e3c7fb9c/volumes" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.632214 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlhp5\" (UniqueName: \"kubernetes.io/projected/69f6622f-9286-49ea-b8b7-f56a61143e03-kube-api-access-dlhp5\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.632289 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.632382 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.632410 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.632506 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.734592 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlhp5\" (UniqueName: \"kubernetes.io/projected/69f6622f-9286-49ea-b8b7-f56a61143e03-kube-api-access-dlhp5\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.734699 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.734839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.734878 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.735013 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.740957 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.740976 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.741371 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.752822 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f6622f-9286-49ea-b8b7-f56a61143e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.753098 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlhp5\" (UniqueName: \"kubernetes.io/projected/69f6622f-9286-49ea-b8b7-f56a61143e03-kube-api-access-dlhp5\") pod \"nova-cell1-novncproxy-0\" (UID: \"69f6622f-9286-49ea-b8b7-f56a61143e03\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:09 crc kubenswrapper[4796]: I1127 11:59:09.859638 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:10 crc kubenswrapper[4796]: I1127 11:59:10.315114 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 11:59:10 crc kubenswrapper[4796]: W1127 11:59:10.316075 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69f6622f_9286_49ea_b8b7_f56a61143e03.slice/crio-e8e0c4d2619f5ddc84425634714d73d76572fa4cf56766846ade737587f2a69a WatchSource:0}: Error finding container e8e0c4d2619f5ddc84425634714d73d76572fa4cf56766846ade737587f2a69a: Status 404 returned error can't find the container with id e8e0c4d2619f5ddc84425634714d73d76572fa4cf56766846ade737587f2a69a Nov 27 11:59:10 crc kubenswrapper[4796]: I1127 11:59:10.465709 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"69f6622f-9286-49ea-b8b7-f56a61143e03","Type":"ContainerStarted","Data":"e8e0c4d2619f5ddc84425634714d73d76572fa4cf56766846ade737587f2a69a"} Nov 27 11:59:11 crc kubenswrapper[4796]: I1127 11:59:11.479353 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"69f6622f-9286-49ea-b8b7-f56a61143e03","Type":"ContainerStarted","Data":"6f971849abfd462ba5a9ef5447f0fed4b972ba3781ba463e6f0ef02ef5e78e01"} Nov 27 11:59:11 crc kubenswrapper[4796]: I1127 11:59:11.499478 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.499461635 podStartE2EDuration="2.499461635s" podCreationTimestamp="2025-11-27 11:59:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:11.497770239 +0000 UTC m=+2069.016089157" watchObservedRunningTime="2025-11-27 11:59:11.499461635 +0000 UTC m=+2069.017780553" Nov 27 11:59:11 crc kubenswrapper[4796]: I1127 11:59:11.503631 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 11:59:11 crc kubenswrapper[4796]: I1127 11:59:11.504739 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 11:59:11 crc kubenswrapper[4796]: I1127 11:59:11.510718 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 11:59:11 crc kubenswrapper[4796]: I1127 11:59:11.517305 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.490639 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.496517 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.561182 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.561495 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c3f46f0d-069e-4527-a8a5-ca60f6022f9c" containerName="kube-state-metrics" containerID="cri-o://2ea395df8fdccd851176858cf9e794dcec4fef06c99a2aa68873cbdbf79f88e5" gracePeriod=30 Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.710712 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-qfbbw"] Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.721305 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.726354 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-qfbbw"] Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.807385 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xshmn\" (UniqueName: \"kubernetes.io/projected/bcb72c79-d085-44f4-96c8-850aaeeb89f4-kube-api-access-xshmn\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.807434 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.807483 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.807507 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-config\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.807524 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.807546 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.908858 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xshmn\" (UniqueName: \"kubernetes.io/projected/bcb72c79-d085-44f4-96c8-850aaeeb89f4-kube-api-access-xshmn\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.908905 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.908947 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.908967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-config\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.908983 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.909009 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.910096 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-config\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.910483 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.910511 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.910996 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.911334 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:12 crc kubenswrapper[4796]: I1127 11:59:12.932904 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xshmn\" (UniqueName: \"kubernetes.io/projected/bcb72c79-d085-44f4-96c8-850aaeeb89f4-kube-api-access-xshmn\") pod \"dnsmasq-dns-f84f9ccf-qfbbw\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:13 crc kubenswrapper[4796]: I1127 11:59:13.122458 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:13 crc kubenswrapper[4796]: I1127 11:59:13.498169 4796 generic.go:334] "Generic (PLEG): container finished" podID="c3f46f0d-069e-4527-a8a5-ca60f6022f9c" containerID="2ea395df8fdccd851176858cf9e794dcec4fef06c99a2aa68873cbdbf79f88e5" exitCode=2 Nov 27 11:59:13 crc kubenswrapper[4796]: I1127 11:59:13.498658 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3f46f0d-069e-4527-a8a5-ca60f6022f9c","Type":"ContainerDied","Data":"2ea395df8fdccd851176858cf9e794dcec4fef06c99a2aa68873cbdbf79f88e5"} Nov 27 11:59:13 crc kubenswrapper[4796]: I1127 11:59:13.617093 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-qfbbw"] Nov 27 11:59:13 crc kubenswrapper[4796]: W1127 11:59:13.646439 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcb72c79_d085_44f4_96c8_850aaeeb89f4.slice/crio-968e866e01e2cdbd68146c79e6eedfaaef22fad96d2b728d25f2e10150b5fbfa WatchSource:0}: Error finding container 968e866e01e2cdbd68146c79e6eedfaaef22fad96d2b728d25f2e10150b5fbfa: Status 404 returned error can't find the container with id 968e866e01e2cdbd68146c79e6eedfaaef22fad96d2b728d25f2e10150b5fbfa Nov 27 11:59:13 crc kubenswrapper[4796]: I1127 11:59:13.884398 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.031934 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twdrc\" (UniqueName: \"kubernetes.io/projected/c3f46f0d-069e-4527-a8a5-ca60f6022f9c-kube-api-access-twdrc\") pod \"c3f46f0d-069e-4527-a8a5-ca60f6022f9c\" (UID: \"c3f46f0d-069e-4527-a8a5-ca60f6022f9c\") " Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.042507 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3f46f0d-069e-4527-a8a5-ca60f6022f9c-kube-api-access-twdrc" (OuterVolumeSpecName: "kube-api-access-twdrc") pod "c3f46f0d-069e-4527-a8a5-ca60f6022f9c" (UID: "c3f46f0d-069e-4527-a8a5-ca60f6022f9c"). InnerVolumeSpecName "kube-api-access-twdrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.133837 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twdrc\" (UniqueName: \"kubernetes.io/projected/c3f46f0d-069e-4527-a8a5-ca60f6022f9c-kube-api-access-twdrc\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.509674 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" event={"ID":"bcb72c79-d085-44f4-96c8-850aaeeb89f4","Type":"ContainerStarted","Data":"9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417"} Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.509989 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" event={"ID":"bcb72c79-d085-44f4-96c8-850aaeeb89f4","Type":"ContainerStarted","Data":"968e866e01e2cdbd68146c79e6eedfaaef22fad96d2b728d25f2e10150b5fbfa"} Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.513627 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3f46f0d-069e-4527-a8a5-ca60f6022f9c","Type":"ContainerDied","Data":"9e67cecbcf04d09da4fe90137c7042fd3a1d52ce4ad34281ff09d1a43871738f"} Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.513661 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.513701 4796 scope.go:117] "RemoveContainer" containerID="2ea395df8fdccd851176858cf9e794dcec4fef06c99a2aa68873cbdbf79f88e5" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.578664 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.591325 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.609152 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:59:14 crc kubenswrapper[4796]: E1127 11:59:14.609886 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f46f0d-069e-4527-a8a5-ca60f6022f9c" containerName="kube-state-metrics" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.609908 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f46f0d-069e-4527-a8a5-ca60f6022f9c" containerName="kube-state-metrics" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.610106 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f46f0d-069e-4527-a8a5-ca60f6022f9c" containerName="kube-state-metrics" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.610845 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.610854 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.618304 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.618935 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.746587 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.746924 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.753074 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4rdh\" (UniqueName: \"kubernetes.io/projected/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-api-access-v4rdh\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.753227 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.763470 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.763776 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-central-agent" containerID="cri-o://835a2be64574b9da32af7034626adf0469099c4ab8a94e726e11b866bc281b24" gracePeriod=30 Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.763912 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="proxy-httpd" containerID="cri-o://f58112ad4949b195b7e25400bb8a0a3889a8420c80260c5293861185fe26ac3d" gracePeriod=30 Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.763958 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="sg-core" containerID="cri-o://661989fb593bf91f4379032d8a1ad87bed6d17f13f04753aefd24449606f5b04" gracePeriod=30 Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.763991 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-notification-agent" containerID="cri-o://1a3dab3268132c4f7eea1f8bcedff7f8cc2481a1e7efb83a060dc17002a040c9" gracePeriod=30 Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.855378 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.855459 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.855481 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.855580 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4rdh\" (UniqueName: \"kubernetes.io/projected/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-api-access-v4rdh\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.859902 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.860126 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.871000 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.871859 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fecd54-d3de-4157-b04e-3187be01c0ad-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.873991 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4rdh\" (UniqueName: \"kubernetes.io/projected/82fecd54-d3de-4157-b04e-3187be01c0ad-kube-api-access-v4rdh\") pod \"kube-state-metrics-0\" (UID: \"82fecd54-d3de-4157-b04e-3187be01c0ad\") " pod="openstack/kube-state-metrics-0" Nov 27 11:59:14 crc kubenswrapper[4796]: I1127 11:59:14.936557 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 11:59:15 crc kubenswrapper[4796]: W1127 11:59:15.417302 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82fecd54_d3de_4157_b04e_3187be01c0ad.slice/crio-76a87e1e250730d6f3a10e3ca77d0699e6e6acdb09553cc7af7edfdde0629203 WatchSource:0}: Error finding container 76a87e1e250730d6f3a10e3ca77d0699e6e6acdb09553cc7af7edfdde0629203: Status 404 returned error can't find the container with id 76a87e1e250730d6f3a10e3ca77d0699e6e6acdb09553cc7af7edfdde0629203 Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.418924 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.473188 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.525239 4796 generic.go:334] "Generic (PLEG): container finished" podID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerID="f58112ad4949b195b7e25400bb8a0a3889a8420c80260c5293861185fe26ac3d" exitCode=0 Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.525288 4796 generic.go:334] "Generic (PLEG): container finished" podID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerID="661989fb593bf91f4379032d8a1ad87bed6d17f13f04753aefd24449606f5b04" exitCode=2 Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.525306 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerDied","Data":"f58112ad4949b195b7e25400bb8a0a3889a8420c80260c5293861185fe26ac3d"} Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.525348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerDied","Data":"661989fb593bf91f4379032d8a1ad87bed6d17f13f04753aefd24449606f5b04"} Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.528909 4796 generic.go:334] "Generic (PLEG): container finished" podID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerID="9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417" exitCode=0 Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.528990 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" event={"ID":"bcb72c79-d085-44f4-96c8-850aaeeb89f4","Type":"ContainerDied","Data":"9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417"} Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.531374 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82fecd54-d3de-4157-b04e-3187be01c0ad","Type":"ContainerStarted","Data":"76a87e1e250730d6f3a10e3ca77d0699e6e6acdb09553cc7af7edfdde0629203"} Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.535588 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-log" containerID="cri-o://82711683561e67c4ef1cbd66ebf23b9f161f6a47d40df81efced6fc104a13cde" gracePeriod=30 Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.535735 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-api" containerID="cri-o://ae0f76141761dc767260ee17ed69999da02ae0f432b6ee873c35f645bcfb025b" gracePeriod=30 Nov 27 11:59:15 crc kubenswrapper[4796]: I1127 11:59:15.586604 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3f46f0d-069e-4527-a8a5-ca60f6022f9c" path="/var/lib/kubelet/pods/c3f46f0d-069e-4527-a8a5-ca60f6022f9c/volumes" Nov 27 11:59:18 crc kubenswrapper[4796]: I1127 11:59:18.568717 4796 generic.go:334] "Generic (PLEG): container finished" podID="3017636c-aebd-45bf-a21c-5cee6583369d" containerID="82711683561e67c4ef1cbd66ebf23b9f161f6a47d40df81efced6fc104a13cde" exitCode=143 Nov 27 11:59:18 crc kubenswrapper[4796]: I1127 11:59:18.570012 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3017636c-aebd-45bf-a21c-5cee6583369d","Type":"ContainerDied","Data":"82711683561e67c4ef1cbd66ebf23b9f161f6a47d40df81efced6fc104a13cde"} Nov 27 11:59:18 crc kubenswrapper[4796]: I1127 11:59:18.572401 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" event={"ID":"bcb72c79-d085-44f4-96c8-850aaeeb89f4","Type":"ContainerStarted","Data":"bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17"} Nov 27 11:59:18 crc kubenswrapper[4796]: I1127 11:59:18.575916 4796 generic.go:334] "Generic (PLEG): container finished" podID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerID="835a2be64574b9da32af7034626adf0469099c4ab8a94e726e11b866bc281b24" exitCode=0 Nov 27 11:59:18 crc kubenswrapper[4796]: I1127 11:59:18.575995 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerDied","Data":"835a2be64574b9da32af7034626adf0469099c4ab8a94e726e11b866bc281b24"} Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.589253 4796 generic.go:334] "Generic (PLEG): container finished" podID="3017636c-aebd-45bf-a21c-5cee6583369d" containerID="ae0f76141761dc767260ee17ed69999da02ae0f432b6ee873c35f645bcfb025b" exitCode=0 Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.589702 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3017636c-aebd-45bf-a21c-5cee6583369d","Type":"ContainerDied","Data":"ae0f76141761dc767260ee17ed69999da02ae0f432b6ee873c35f645bcfb025b"} Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.596077 4796 generic.go:334] "Generic (PLEG): container finished" podID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerID="1a3dab3268132c4f7eea1f8bcedff7f8cc2481a1e7efb83a060dc17002a040c9" exitCode=0 Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.596166 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerDied","Data":"1a3dab3268132c4f7eea1f8bcedff7f8cc2481a1e7efb83a060dc17002a040c9"} Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.596332 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.627114 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" podStartSLOduration=7.627096016 podStartE2EDuration="7.627096016s" podCreationTimestamp="2025-11-27 11:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:19.615585772 +0000 UTC m=+2077.133904690" watchObservedRunningTime="2025-11-27 11:59:19.627096016 +0000 UTC m=+2077.145414944" Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.860395 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:19 crc kubenswrapper[4796]: I1127 11:59:19.917739 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.054305 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.187654 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-config-data\") pod \"3017636c-aebd-45bf-a21c-5cee6583369d\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.188073 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3017636c-aebd-45bf-a21c-5cee6583369d-logs\") pod \"3017636c-aebd-45bf-a21c-5cee6583369d\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.188159 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cljp2\" (UniqueName: \"kubernetes.io/projected/3017636c-aebd-45bf-a21c-5cee6583369d-kube-api-access-cljp2\") pod \"3017636c-aebd-45bf-a21c-5cee6583369d\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.189227 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3017636c-aebd-45bf-a21c-5cee6583369d-logs" (OuterVolumeSpecName: "logs") pod "3017636c-aebd-45bf-a21c-5cee6583369d" (UID: "3017636c-aebd-45bf-a21c-5cee6583369d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.189747 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-combined-ca-bundle\") pod \"3017636c-aebd-45bf-a21c-5cee6583369d\" (UID: \"3017636c-aebd-45bf-a21c-5cee6583369d\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.190743 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3017636c-aebd-45bf-a21c-5cee6583369d-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.195680 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3017636c-aebd-45bf-a21c-5cee6583369d-kube-api-access-cljp2" (OuterVolumeSpecName: "kube-api-access-cljp2") pod "3017636c-aebd-45bf-a21c-5cee6583369d" (UID: "3017636c-aebd-45bf-a21c-5cee6583369d"). InnerVolumeSpecName "kube-api-access-cljp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.231339 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3017636c-aebd-45bf-a21c-5cee6583369d" (UID: "3017636c-aebd-45bf-a21c-5cee6583369d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.233369 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-config-data" (OuterVolumeSpecName: "config-data") pod "3017636c-aebd-45bf-a21c-5cee6583369d" (UID: "3017636c-aebd-45bf-a21c-5cee6583369d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.292067 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.292095 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3017636c-aebd-45bf-a21c-5cee6583369d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.292103 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cljp2\" (UniqueName: \"kubernetes.io/projected/3017636c-aebd-45bf-a21c-5cee6583369d-kube-api-access-cljp2\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.378437 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.494781 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8kbd\" (UniqueName: \"kubernetes.io/projected/522aafbf-5229-44e5-b61c-d194b5aca60b-kube-api-access-n8kbd\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.494956 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-combined-ca-bundle\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.495018 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-log-httpd\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.495037 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-scripts\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.495052 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-sg-core-conf-yaml\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.495074 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-config-data\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.495106 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-run-httpd\") pod \"522aafbf-5229-44e5-b61c-d194b5aca60b\" (UID: \"522aafbf-5229-44e5-b61c-d194b5aca60b\") " Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.496035 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.497720 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.497898 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-scripts" (OuterVolumeSpecName: "scripts") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.498243 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522aafbf-5229-44e5-b61c-d194b5aca60b-kube-api-access-n8kbd" (OuterVolumeSpecName: "kube-api-access-n8kbd") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "kube-api-access-n8kbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.523586 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.570426 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.597059 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.597089 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.597099 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.597114 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.597123 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/522aafbf-5229-44e5-b61c-d194b5aca60b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.597131 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8kbd\" (UniqueName: \"kubernetes.io/projected/522aafbf-5229-44e5-b61c-d194b5aca60b-kube-api-access-n8kbd\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.611772 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82fecd54-d3de-4157-b04e-3187be01c0ad","Type":"ContainerStarted","Data":"5e693973484b9528e6cf27789285014ebc9026c7e4bdddb0716c7ad8c2a7c9e4"} Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.612495 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.612925 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-config-data" (OuterVolumeSpecName: "config-data") pod "522aafbf-5229-44e5-b61c-d194b5aca60b" (UID: "522aafbf-5229-44e5-b61c-d194b5aca60b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.615448 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"522aafbf-5229-44e5-b61c-d194b5aca60b","Type":"ContainerDied","Data":"64c9ee311f83df9fe0b71c1ba7ad9318d33773fc07ca4b183265d87e420bbeda"} Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.615505 4796 scope.go:117] "RemoveContainer" containerID="f58112ad4949b195b7e25400bb8a0a3889a8420c80260c5293861185fe26ac3d" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.615598 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.622993 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.623060 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3017636c-aebd-45bf-a21c-5cee6583369d","Type":"ContainerDied","Data":"e5fa804ed46c13507e4aac78a587d093b4aca82698e97ef9b12ed461e0d6ac4b"} Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.631133 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.86615848 podStartE2EDuration="6.631109602s" podCreationTimestamp="2025-11-27 11:59:14 +0000 UTC" firstStartedPulling="2025-11-27 11:59:15.420464732 +0000 UTC m=+2072.938783650" lastFinishedPulling="2025-11-27 11:59:20.185415854 +0000 UTC m=+2077.703734772" observedRunningTime="2025-11-27 11:59:20.629735765 +0000 UTC m=+2078.148054683" watchObservedRunningTime="2025-11-27 11:59:20.631109602 +0000 UTC m=+2078.149428520" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.643292 4796 scope.go:117] "RemoveContainer" containerID="661989fb593bf91f4379032d8a1ad87bed6d17f13f04753aefd24449606f5b04" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.645570 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.663687 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.677869 4796 scope.go:117] "RemoveContainer" containerID="1a3dab3268132c4f7eea1f8bcedff7f8cc2481a1e7efb83a060dc17002a040c9" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.689377 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.699397 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522aafbf-5229-44e5-b61c-d194b5aca60b-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.718744 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.726386 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.742707 4796 scope.go:117] "RemoveContainer" containerID="835a2be64574b9da32af7034626adf0469099c4ab8a94e726e11b866bc281b24" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745203 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: E1127 11:59:20.745694 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-central-agent" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745720 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-central-agent" Nov 27 11:59:20 crc kubenswrapper[4796]: E1127 11:59:20.745746 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="sg-core" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745755 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="sg-core" Nov 27 11:59:20 crc kubenswrapper[4796]: E1127 11:59:20.745772 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-api" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745781 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-api" Nov 27 11:59:20 crc kubenswrapper[4796]: E1127 11:59:20.745798 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-notification-agent" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745806 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-notification-agent" Nov 27 11:59:20 crc kubenswrapper[4796]: E1127 11:59:20.745819 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-log" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745827 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-log" Nov 27 11:59:20 crc kubenswrapper[4796]: E1127 11:59:20.745860 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="proxy-httpd" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.745870 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="proxy-httpd" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.746082 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-api" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.746097 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="proxy-httpd" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.746108 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-central-agent" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.746132 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="ceilometer-notification-agent" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.746146 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" containerName="nova-api-log" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.746161 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" containerName="sg-core" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.748257 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.750849 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.751011 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.751009 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.758013 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.783725 4796 scope.go:117] "RemoveContainer" containerID="ae0f76141761dc767260ee17ed69999da02ae0f432b6ee873c35f645bcfb025b" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.783858 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.785857 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.789891 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.789980 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.790624 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.793220 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.847246 4796 scope.go:117] "RemoveContainer" containerID="82711683561e67c4ef1cbd66ebf23b9f161f6a47d40df81efced6fc104a13cde" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.872128 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-srjn9"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.873365 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.877377 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.877478 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.883778 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-srjn9"] Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904235 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-run-httpd\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904305 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904353 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904503 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a33aa29-4670-4c44-bbae-428de553ca8f-logs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-config-data\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904645 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-scripts\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904669 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqlh9\" (UniqueName: \"kubernetes.io/projected/717da22f-c39f-46f3-bd9f-437f6becf5c4-kube-api-access-dqlh9\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904690 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4fjn\" (UniqueName: \"kubernetes.io/projected/4a33aa29-4670-4c44-bbae-428de553ca8f-kube-api-access-r4fjn\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904840 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-log-httpd\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.904918 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.905005 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-public-tls-certs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.905082 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.905118 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:20 crc kubenswrapper[4796]: I1127 11:59:20.905195 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-config-data\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.006812 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-config-data\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.006879 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-scripts\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.006929 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqlh9\" (UniqueName: \"kubernetes.io/projected/717da22f-c39f-46f3-bd9f-437f6becf5c4-kube-api-access-dqlh9\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.006950 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4fjn\" (UniqueName: \"kubernetes.io/projected/4a33aa29-4670-4c44-bbae-428de553ca8f-kube-api-access-r4fjn\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007614 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007653 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-log-httpd\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007687 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkwsd\" (UniqueName: \"kubernetes.io/projected/5d77f0e3-d2ec-485a-9616-a2fde63f185a-kube-api-access-gkwsd\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007704 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007722 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-public-tls-certs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007782 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007872 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-config-data\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.007970 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-scripts\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008042 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-run-httpd\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008118 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008147 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-log-httpd\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008145 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008249 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-config-data\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008291 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a33aa29-4670-4c44-bbae-428de553ca8f-logs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008351 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-run-httpd\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.008724 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a33aa29-4670-4c44-bbae-428de553ca8f-logs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.011724 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.012844 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-config-data\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.012949 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-public-tls-certs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.012961 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-scripts\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.013445 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-config-data\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.013694 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.014374 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.014992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.016089 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.023945 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqlh9\" (UniqueName: \"kubernetes.io/projected/717da22f-c39f-46f3-bd9f-437f6becf5c4-kube-api-access-dqlh9\") pod \"ceilometer-0\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.028796 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4fjn\" (UniqueName: \"kubernetes.io/projected/4a33aa29-4670-4c44-bbae-428de553ca8f-kube-api-access-r4fjn\") pod \"nova-api-0\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.084006 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.109648 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-config-data\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.109740 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.109766 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkwsd\" (UniqueName: \"kubernetes.io/projected/5d77f0e3-d2ec-485a-9616-a2fde63f185a-kube-api-access-gkwsd\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.109838 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-scripts\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.115064 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-config-data\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.115191 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.115453 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.116051 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-scripts\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.128389 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkwsd\" (UniqueName: \"kubernetes.io/projected/5d77f0e3-d2ec-485a-9616-a2fde63f185a-kube-api-access-gkwsd\") pod \"nova-cell1-cell-mapping-srjn9\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.194964 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.580305 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3017636c-aebd-45bf-a21c-5cee6583369d" path="/var/lib/kubelet/pods/3017636c-aebd-45bf-a21c-5cee6583369d/volumes" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.581070 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522aafbf-5229-44e5-b61c-d194b5aca60b" path="/var/lib/kubelet/pods/522aafbf-5229-44e5-b61c-d194b5aca60b/volumes" Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.612167 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:21 crc kubenswrapper[4796]: W1127 11:59:21.619789 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod717da22f_c39f_46f3_bd9f_437f6becf5c4.slice/crio-3ea3e82636d1633e96b20973349c2d4c4d9826f7449167fa97705636b991990d WatchSource:0}: Error finding container 3ea3e82636d1633e96b20973349c2d4c4d9826f7449167fa97705636b991990d: Status 404 returned error can't find the container with id 3ea3e82636d1633e96b20973349c2d4c4d9826f7449167fa97705636b991990d Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.634427 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerStarted","Data":"3ea3e82636d1633e96b20973349c2d4c4d9826f7449167fa97705636b991990d"} Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.737036 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.831634 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-srjn9"] Nov 27 11:59:21 crc kubenswrapper[4796]: W1127 11:59:21.836789 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d77f0e3_d2ec_485a_9616_a2fde63f185a.slice/crio-44650bd57b5da97597f72d8acc204a0abfd978368b839306902b4d9e4be0fa81 WatchSource:0}: Error finding container 44650bd57b5da97597f72d8acc204a0abfd978368b839306902b4d9e4be0fa81: Status 404 returned error can't find the container with id 44650bd57b5da97597f72d8acc204a0abfd978368b839306902b4d9e4be0fa81 Nov 27 11:59:21 crc kubenswrapper[4796]: I1127 11:59:21.893542 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.650479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a33aa29-4670-4c44-bbae-428de553ca8f","Type":"ContainerStarted","Data":"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a"} Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.650823 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a33aa29-4670-4c44-bbae-428de553ca8f","Type":"ContainerStarted","Data":"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3"} Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.650833 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a33aa29-4670-4c44-bbae-428de553ca8f","Type":"ContainerStarted","Data":"7e8afec2f6e29553a5afe31be867210b829ac3e01417bef6c38fe931627accdf"} Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.671254 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-srjn9" event={"ID":"5d77f0e3-d2ec-485a-9616-a2fde63f185a","Type":"ContainerStarted","Data":"c500059c4dfad83633a876b9625956d648e9e712c9d714223b95c3a9dff4e27e"} Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.671329 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-srjn9" event={"ID":"5d77f0e3-d2ec-485a-9616-a2fde63f185a","Type":"ContainerStarted","Data":"44650bd57b5da97597f72d8acc204a0abfd978368b839306902b4d9e4be0fa81"} Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.698540 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.698519075 podStartE2EDuration="2.698519075s" podCreationTimestamp="2025-11-27 11:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:22.696860411 +0000 UTC m=+2080.215179329" watchObservedRunningTime="2025-11-27 11:59:22.698519075 +0000 UTC m=+2080.216837993" Nov 27 11:59:22 crc kubenswrapper[4796]: I1127 11:59:22.724119 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-srjn9" podStartSLOduration=2.724097683 podStartE2EDuration="2.724097683s" podCreationTimestamp="2025-11-27 11:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:22.717354579 +0000 UTC m=+2080.235673497" watchObservedRunningTime="2025-11-27 11:59:22.724097683 +0000 UTC m=+2080.242416601" Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.124336 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.213570 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-kg5xv"] Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.213815 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerName="dnsmasq-dns" containerID="cri-o://7bf76cf318fe6e654a801a4bf85ab7f81f2e81f0051a883e87825affca93c2ca" gracePeriod=10 Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.684511 4796 generic.go:334] "Generic (PLEG): container finished" podID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerID="7bf76cf318fe6e654a801a4bf85ab7f81f2e81f0051a883e87825affca93c2ca" exitCode=0 Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.684602 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" event={"ID":"e17ae35d-43d7-4afb-947e-db43516a6a92","Type":"ContainerDied","Data":"7bf76cf318fe6e654a801a4bf85ab7f81f2e81f0051a883e87825affca93c2ca"} Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.687106 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerStarted","Data":"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664"} Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.841056 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.974244 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnsqq\" (UniqueName: \"kubernetes.io/projected/e17ae35d-43d7-4afb-947e-db43516a6a92-kube-api-access-xnsqq\") pod \"e17ae35d-43d7-4afb-947e-db43516a6a92\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.974350 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-nb\") pod \"e17ae35d-43d7-4afb-947e-db43516a6a92\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.974391 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-swift-storage-0\") pod \"e17ae35d-43d7-4afb-947e-db43516a6a92\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.974443 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-config\") pod \"e17ae35d-43d7-4afb-947e-db43516a6a92\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.974513 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-sb\") pod \"e17ae35d-43d7-4afb-947e-db43516a6a92\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.974543 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-svc\") pod \"e17ae35d-43d7-4afb-947e-db43516a6a92\" (UID: \"e17ae35d-43d7-4afb-947e-db43516a6a92\") " Nov 27 11:59:23 crc kubenswrapper[4796]: I1127 11:59:23.981540 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e17ae35d-43d7-4afb-947e-db43516a6a92-kube-api-access-xnsqq" (OuterVolumeSpecName: "kube-api-access-xnsqq") pod "e17ae35d-43d7-4afb-947e-db43516a6a92" (UID: "e17ae35d-43d7-4afb-947e-db43516a6a92"). InnerVolumeSpecName "kube-api-access-xnsqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.032768 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e17ae35d-43d7-4afb-947e-db43516a6a92" (UID: "e17ae35d-43d7-4afb-947e-db43516a6a92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.033259 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e17ae35d-43d7-4afb-947e-db43516a6a92" (UID: "e17ae35d-43d7-4afb-947e-db43516a6a92"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.033603 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-config" (OuterVolumeSpecName: "config") pod "e17ae35d-43d7-4afb-947e-db43516a6a92" (UID: "e17ae35d-43d7-4afb-947e-db43516a6a92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.041666 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e17ae35d-43d7-4afb-947e-db43516a6a92" (UID: "e17ae35d-43d7-4afb-947e-db43516a6a92"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.045712 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e17ae35d-43d7-4afb-947e-db43516a6a92" (UID: "e17ae35d-43d7-4afb-947e-db43516a6a92"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.077097 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnsqq\" (UniqueName: \"kubernetes.io/projected/e17ae35d-43d7-4afb-947e-db43516a6a92-kube-api-access-xnsqq\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.077143 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.077154 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.077165 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-config\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.077173 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.077183 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e17ae35d-43d7-4afb-947e-db43516a6a92-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.702307 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" event={"ID":"e17ae35d-43d7-4afb-947e-db43516a6a92","Type":"ContainerDied","Data":"b436f7d479cee8ea816572ad754051293ee4f59aff87723b2e054dddedb0c90c"} Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.702888 4796 scope.go:117] "RemoveContainer" containerID="7bf76cf318fe6e654a801a4bf85ab7f81f2e81f0051a883e87825affca93c2ca" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.702692 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-kg5xv" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.754284 4796 scope.go:117] "RemoveContainer" containerID="04ed696bf6ee51aec2ae57cb48a8d6f078a90330b94ad247b1e0c165c899004e" Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.787703 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-kg5xv"] Nov 27 11:59:24 crc kubenswrapper[4796]: I1127 11:59:24.811647 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-kg5xv"] Nov 27 11:59:25 crc kubenswrapper[4796]: I1127 11:59:25.582442 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" path="/var/lib/kubelet/pods/e17ae35d-43d7-4afb-947e-db43516a6a92/volumes" Nov 27 11:59:25 crc kubenswrapper[4796]: I1127 11:59:25.714244 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerStarted","Data":"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b"} Nov 27 11:59:26 crc kubenswrapper[4796]: I1127 11:59:26.751861 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerStarted","Data":"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617"} Nov 27 11:59:28 crc kubenswrapper[4796]: I1127 11:59:28.791052 4796 generic.go:334] "Generic (PLEG): container finished" podID="5d77f0e3-d2ec-485a-9616-a2fde63f185a" containerID="c500059c4dfad83633a876b9625956d648e9e712c9d714223b95c3a9dff4e27e" exitCode=0 Nov 27 11:59:28 crc kubenswrapper[4796]: I1127 11:59:28.791157 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-srjn9" event={"ID":"5d77f0e3-d2ec-485a-9616-a2fde63f185a","Type":"ContainerDied","Data":"c500059c4dfad83633a876b9625956d648e9e712c9d714223b95c3a9dff4e27e"} Nov 27 11:59:29 crc kubenswrapper[4796]: I1127 11:59:29.804538 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerStarted","Data":"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6"} Nov 27 11:59:29 crc kubenswrapper[4796]: I1127 11:59:29.804840 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-central-agent" containerID="cri-o://62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" gracePeriod=30 Nov 27 11:59:29 crc kubenswrapper[4796]: I1127 11:59:29.804894 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-notification-agent" containerID="cri-o://1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" gracePeriod=30 Nov 27 11:59:29 crc kubenswrapper[4796]: I1127 11:59:29.804877 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="sg-core" containerID="cri-o://ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" gracePeriod=30 Nov 27 11:59:29 crc kubenswrapper[4796]: I1127 11:59:29.804992 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="proxy-httpd" containerID="cri-o://921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" gracePeriod=30 Nov 27 11:59:29 crc kubenswrapper[4796]: I1127 11:59:29.840571 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.662134865 podStartE2EDuration="9.840551361s" podCreationTimestamp="2025-11-27 11:59:20 +0000 UTC" firstStartedPulling="2025-11-27 11:59:21.624176892 +0000 UTC m=+2079.142495810" lastFinishedPulling="2025-11-27 11:59:28.802593378 +0000 UTC m=+2086.320912306" observedRunningTime="2025-11-27 11:59:29.839025159 +0000 UTC m=+2087.357344087" watchObservedRunningTime="2025-11-27 11:59:29.840551361 +0000 UTC m=+2087.358870279" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.318092 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.424180 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-scripts\") pod \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.424365 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-config-data\") pod \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.424474 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkwsd\" (UniqueName: \"kubernetes.io/projected/5d77f0e3-d2ec-485a-9616-a2fde63f185a-kube-api-access-gkwsd\") pod \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.424525 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-combined-ca-bundle\") pod \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\" (UID: \"5d77f0e3-d2ec-485a-9616-a2fde63f185a\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.430822 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d77f0e3-d2ec-485a-9616-a2fde63f185a-kube-api-access-gkwsd" (OuterVolumeSpecName: "kube-api-access-gkwsd") pod "5d77f0e3-d2ec-485a-9616-a2fde63f185a" (UID: "5d77f0e3-d2ec-485a-9616-a2fde63f185a"). InnerVolumeSpecName "kube-api-access-gkwsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.431334 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-scripts" (OuterVolumeSpecName: "scripts") pod "5d77f0e3-d2ec-485a-9616-a2fde63f185a" (UID: "5d77f0e3-d2ec-485a-9616-a2fde63f185a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.458964 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d77f0e3-d2ec-485a-9616-a2fde63f185a" (UID: "5d77f0e3-d2ec-485a-9616-a2fde63f185a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.465869 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-config-data" (OuterVolumeSpecName: "config-data") pod "5d77f0e3-d2ec-485a-9616-a2fde63f185a" (UID: "5d77f0e3-d2ec-485a-9616-a2fde63f185a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.489566 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.527054 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.527108 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkwsd\" (UniqueName: \"kubernetes.io/projected/5d77f0e3-d2ec-485a-9616-a2fde63f185a-kube-api-access-gkwsd\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.527127 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.527145 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d77f0e3-d2ec-485a-9616-a2fde63f185a-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628431 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqlh9\" (UniqueName: \"kubernetes.io/projected/717da22f-c39f-46f3-bd9f-437f6becf5c4-kube-api-access-dqlh9\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628572 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-config-data\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628774 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-scripts\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628816 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-sg-core-conf-yaml\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628865 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-combined-ca-bundle\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628893 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-log-httpd\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628932 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-run-httpd\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.628993 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-ceilometer-tls-certs\") pod \"717da22f-c39f-46f3-bd9f-437f6becf5c4\" (UID: \"717da22f-c39f-46f3-bd9f-437f6becf5c4\") " Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.629860 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.631595 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.632049 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717da22f-c39f-46f3-bd9f-437f6becf5c4-kube-api-access-dqlh9" (OuterVolumeSpecName: "kube-api-access-dqlh9") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "kube-api-access-dqlh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.635208 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-scripts" (OuterVolumeSpecName: "scripts") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.655376 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.711917 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.724471 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732121 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732409 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732570 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732685 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/717da22f-c39f-46f3-bd9f-437f6becf5c4-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732765 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732851 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqlh9\" (UniqueName: \"kubernetes.io/projected/717da22f-c39f-46f3-bd9f-437f6becf5c4-kube-api-access-dqlh9\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.732950 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.740239 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-config-data" (OuterVolumeSpecName: "config-data") pod "717da22f-c39f-46f3-bd9f-437f6becf5c4" (UID: "717da22f-c39f-46f3-bd9f-437f6becf5c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.817903 4796 generic.go:334] "Generic (PLEG): container finished" podID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" exitCode=0 Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.817955 4796 generic.go:334] "Generic (PLEG): container finished" podID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" exitCode=2 Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.817973 4796 generic.go:334] "Generic (PLEG): container finished" podID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" exitCode=0 Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.817987 4796 generic.go:334] "Generic (PLEG): container finished" podID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" exitCode=0 Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818019 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerDied","Data":"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6"} Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818645 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerDied","Data":"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617"} Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818673 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerDied","Data":"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b"} Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818690 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerDied","Data":"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664"} Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818709 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"717da22f-c39f-46f3-bd9f-437f6becf5c4","Type":"ContainerDied","Data":"3ea3e82636d1633e96b20973349c2d4c4d9826f7449167fa97705636b991990d"} Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.818734 4796 scope.go:117] "RemoveContainer" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.822853 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-srjn9" event={"ID":"5d77f0e3-d2ec-485a-9616-a2fde63f185a","Type":"ContainerDied","Data":"44650bd57b5da97597f72d8acc204a0abfd978368b839306902b4d9e4be0fa81"} Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.822904 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44650bd57b5da97597f72d8acc204a0abfd978368b839306902b4d9e4be0fa81" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.822936 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-srjn9" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.835913 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/717da22f-c39f-46f3-bd9f-437f6becf5c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.850632 4796 scope.go:117] "RemoveContainer" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.885645 4796 scope.go:117] "RemoveContainer" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.893649 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.914545 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.916291 4796 scope.go:117] "RemoveContainer" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931141 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931593 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-central-agent" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931612 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-central-agent" Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931643 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="sg-core" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931650 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="sg-core" Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931662 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerName="init" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931667 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerName="init" Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931684 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d77f0e3-d2ec-485a-9616-a2fde63f185a" containerName="nova-manage" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931690 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d77f0e3-d2ec-485a-9616-a2fde63f185a" containerName="nova-manage" Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931707 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerName="dnsmasq-dns" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931713 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerName="dnsmasq-dns" Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931724 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-notification-agent" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931730 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-notification-agent" Nov 27 11:59:30 crc kubenswrapper[4796]: E1127 11:59:30.931742 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="proxy-httpd" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931748 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="proxy-httpd" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931916 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-central-agent" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931926 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="proxy-httpd" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931934 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="ceilometer-notification-agent" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931946 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d77f0e3-d2ec-485a-9616-a2fde63f185a" containerName="nova-manage" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931956 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" containerName="sg-core" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.931967 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e17ae35d-43d7-4afb-947e-db43516a6a92" containerName="dnsmasq-dns" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.933644 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.936986 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.937640 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.937975 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 11:59:30 crc kubenswrapper[4796]: I1127 11:59:30.947096 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.029369 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.029622 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-log" containerID="cri-o://8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3" gracePeriod=30 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.030117 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-api" containerID="cri-o://a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a" gracePeriod=30 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043609 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043651 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crcj\" (UniqueName: \"kubernetes.io/projected/afc78f25-c3cb-4627-be10-a6838c84d61f-kube-api-access-2crcj\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043669 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043695 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-scripts\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043763 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-log-httpd\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043788 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043815 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-config-data\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.043856 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-run-httpd\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.051788 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.052051 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="406db992-3094-4324-a2fb-ed288d3ee288" containerName="nova-scheduler-scheduler" containerID="cri-o://df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d" gracePeriod=30 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.052570 4796 scope.go:117] "RemoveContainer" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.053596 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": container with ID starting with 921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6 not found: ID does not exist" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.053659 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6"} err="failed to get container status \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": rpc error: code = NotFound desc = could not find container \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": container with ID starting with 921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.053693 4796 scope.go:117] "RemoveContainer" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.054332 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": container with ID starting with ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617 not found: ID does not exist" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.054649 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617"} err="failed to get container status \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": rpc error: code = NotFound desc = could not find container \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": container with ID starting with ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.054681 4796 scope.go:117] "RemoveContainer" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.056488 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": container with ID starting with 1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b not found: ID does not exist" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.056521 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b"} err="failed to get container status \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": rpc error: code = NotFound desc = could not find container \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": container with ID starting with 1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.056542 4796 scope.go:117] "RemoveContainer" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.056754 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": container with ID starting with 62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664 not found: ID does not exist" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.056779 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664"} err="failed to get container status \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": rpc error: code = NotFound desc = could not find container \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": container with ID starting with 62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.056792 4796 scope.go:117] "RemoveContainer" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.057771 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6"} err="failed to get container status \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": rpc error: code = NotFound desc = could not find container \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": container with ID starting with 921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.057792 4796 scope.go:117] "RemoveContainer" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.058042 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617"} err="failed to get container status \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": rpc error: code = NotFound desc = could not find container \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": container with ID starting with ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.058062 4796 scope.go:117] "RemoveContainer" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.058274 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b"} err="failed to get container status \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": rpc error: code = NotFound desc = could not find container \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": container with ID starting with 1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.058294 4796 scope.go:117] "RemoveContainer" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.058477 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664"} err="failed to get container status \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": rpc error: code = NotFound desc = could not find container \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": container with ID starting with 62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.058497 4796 scope.go:117] "RemoveContainer" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.061443 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6"} err="failed to get container status \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": rpc error: code = NotFound desc = could not find container \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": container with ID starting with 921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.061492 4796 scope.go:117] "RemoveContainer" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.061795 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617"} err="failed to get container status \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": rpc error: code = NotFound desc = could not find container \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": container with ID starting with ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.061822 4796 scope.go:117] "RemoveContainer" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062039 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b"} err="failed to get container status \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": rpc error: code = NotFound desc = could not find container \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": container with ID starting with 1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062069 4796 scope.go:117] "RemoveContainer" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062297 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664"} err="failed to get container status \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": rpc error: code = NotFound desc = could not find container \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": container with ID starting with 62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062320 4796 scope.go:117] "RemoveContainer" containerID="921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062579 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6"} err="failed to get container status \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": rpc error: code = NotFound desc = could not find container \"921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6\": container with ID starting with 921dd18698108244fde1e1595a22953110ee3dc4901ce5bf2495b6e98ccdbdb6 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062606 4796 scope.go:117] "RemoveContainer" containerID="ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062835 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617"} err="failed to get container status \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": rpc error: code = NotFound desc = could not find container \"ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617\": container with ID starting with ad01d63d05706fc8347c478c052d331651a40750875bf93f7cc5746a029cd617 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.062858 4796 scope.go:117] "RemoveContainer" containerID="1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.063075 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b"} err="failed to get container status \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": rpc error: code = NotFound desc = could not find container \"1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b\": container with ID starting with 1ebe586e83088fd4c01537bf14b678781e1d340c0aa5df66a2dab8cc3bedde9b not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.063116 4796 scope.go:117] "RemoveContainer" containerID="62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.064406 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.064543 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664"} err="failed to get container status \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": rpc error: code = NotFound desc = could not find container \"62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664\": container with ID starting with 62378b288a7b6570319c7a757f92169d0181e152c7acea5c6207afbf7678b664 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.064633 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-log" containerID="cri-o://796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254" gracePeriod=30 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.064770 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-metadata" containerID="cri-o://83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026" gracePeriod=30 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148237 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-log-httpd\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148328 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148373 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-config-data\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148421 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-run-httpd\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148522 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148548 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crcj\" (UniqueName: \"kubernetes.io/projected/afc78f25-c3cb-4627-be10-a6838c84d61f-kube-api-access-2crcj\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148569 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148604 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-scripts\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.148989 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-log-httpd\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.152122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-run-httpd\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.155644 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.159874 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.160970 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.161166 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-config-data\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.172052 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-scripts\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.178881 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crcj\" (UniqueName: \"kubernetes.io/projected/afc78f25-c3cb-4627-be10-a6838c84d61f-kube-api-access-2crcj\") pod \"ceilometer-0\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.358499 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.524584 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.582114 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="717da22f-c39f-46f3-bd9f-437f6becf5c4" path="/var/lib/kubelet/pods/717da22f-c39f-46f3-bd9f-437f6becf5c4/volumes" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.660457 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-config-data\") pod \"4a33aa29-4670-4c44-bbae-428de553ca8f\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.660581 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-internal-tls-certs\") pod \"4a33aa29-4670-4c44-bbae-428de553ca8f\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.660654 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-public-tls-certs\") pod \"4a33aa29-4670-4c44-bbae-428de553ca8f\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.660717 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a33aa29-4670-4c44-bbae-428de553ca8f-logs\") pod \"4a33aa29-4670-4c44-bbae-428de553ca8f\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.660767 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-combined-ca-bundle\") pod \"4a33aa29-4670-4c44-bbae-428de553ca8f\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.660843 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4fjn\" (UniqueName: \"kubernetes.io/projected/4a33aa29-4670-4c44-bbae-428de553ca8f-kube-api-access-r4fjn\") pod \"4a33aa29-4670-4c44-bbae-428de553ca8f\" (UID: \"4a33aa29-4670-4c44-bbae-428de553ca8f\") " Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.661260 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a33aa29-4670-4c44-bbae-428de553ca8f-logs" (OuterVolumeSpecName: "logs") pod "4a33aa29-4670-4c44-bbae-428de553ca8f" (UID: "4a33aa29-4670-4c44-bbae-428de553ca8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.661534 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a33aa29-4670-4c44-bbae-428de553ca8f-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.667028 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a33aa29-4670-4c44-bbae-428de553ca8f-kube-api-access-r4fjn" (OuterVolumeSpecName: "kube-api-access-r4fjn") pod "4a33aa29-4670-4c44-bbae-428de553ca8f" (UID: "4a33aa29-4670-4c44-bbae-428de553ca8f"). InnerVolumeSpecName "kube-api-access-r4fjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.689464 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-config-data" (OuterVolumeSpecName: "config-data") pod "4a33aa29-4670-4c44-bbae-428de553ca8f" (UID: "4a33aa29-4670-4c44-bbae-428de553ca8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.705552 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a33aa29-4670-4c44-bbae-428de553ca8f" (UID: "4a33aa29-4670-4c44-bbae-428de553ca8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.719147 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4a33aa29-4670-4c44-bbae-428de553ca8f" (UID: "4a33aa29-4670-4c44-bbae-428de553ca8f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.726570 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4a33aa29-4670-4c44-bbae-428de553ca8f" (UID: "4a33aa29-4670-4c44-bbae-428de553ca8f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.745091 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.747099 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.749533 4796 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.749575 4796 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="406db992-3094-4324-a2fb-ed288d3ee288" containerName="nova-scheduler-scheduler" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.768572 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.768624 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4fjn\" (UniqueName: \"kubernetes.io/projected/4a33aa29-4670-4c44-bbae-428de553ca8f-kube-api-access-r4fjn\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.768642 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.768657 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.768670 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a33aa29-4670-4c44-bbae-428de553ca8f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838314 4796 generic.go:334] "Generic (PLEG): container finished" podID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerID="a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a" exitCode=0 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838373 4796 generic.go:334] "Generic (PLEG): container finished" podID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerID="8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3" exitCode=143 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838409 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838439 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a33aa29-4670-4c44-bbae-428de553ca8f","Type":"ContainerDied","Data":"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a"} Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838472 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a33aa29-4670-4c44-bbae-428de553ca8f","Type":"ContainerDied","Data":"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3"} Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838485 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a33aa29-4670-4c44-bbae-428de553ca8f","Type":"ContainerDied","Data":"7e8afec2f6e29553a5afe31be867210b829ac3e01417bef6c38fe931627accdf"} Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.838504 4796 scope.go:117] "RemoveContainer" containerID="a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.840938 4796 generic.go:334] "Generic (PLEG): container finished" podID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerID="796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254" exitCode=143 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.841102 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5a5c6984-1064-4944-a335-9cd8fa5c1738","Type":"ContainerDied","Data":"796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254"} Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.873242 4796 scope.go:117] "RemoveContainer" containerID="8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.877424 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: W1127 11:59:31.879827 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafc78f25_c3cb_4627_be10_a6838c84d61f.slice/crio-55c2f0680d2d54bb4f7fb83c6c998b02b8141f5e0f2f28fab4d35662bb70c576 WatchSource:0}: Error finding container 55c2f0680d2d54bb4f7fb83c6c998b02b8141f5e0f2f28fab4d35662bb70c576: Status 404 returned error can't find the container with id 55c2f0680d2d54bb4f7fb83c6c998b02b8141f5e0f2f28fab4d35662bb70c576 Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.893330 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.899468 4796 scope.go:117] "RemoveContainer" containerID="a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.903826 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a\": container with ID starting with a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a not found: ID does not exist" containerID="a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.904055 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a"} err="failed to get container status \"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a\": rpc error: code = NotFound desc = could not find container \"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a\": container with ID starting with a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.904081 4796 scope.go:117] "RemoveContainer" containerID="8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.904586 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3\": container with ID starting with 8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3 not found: ID does not exist" containerID="8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.904611 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3"} err="failed to get container status \"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3\": rpc error: code = NotFound desc = could not find container \"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3\": container with ID starting with 8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.904626 4796 scope.go:117] "RemoveContainer" containerID="a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.904926 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a"} err="failed to get container status \"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a\": rpc error: code = NotFound desc = could not find container \"a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a\": container with ID starting with a96c4069af59cca94ff03211ba66ff9f5e66418d9bed925251dff8706d549e7a not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.904969 4796 scope.go:117] "RemoveContainer" containerID="8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.905212 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3"} err="failed to get container status \"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3\": rpc error: code = NotFound desc = could not find container \"8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3\": container with ID starting with 8b2bef31d4440867a8671e88a8c3417695db537d6e3c83a8896c8fdae98337a3 not found: ID does not exist" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.908235 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.926454 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.928332 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-log" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.928486 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-log" Nov 27 11:59:31 crc kubenswrapper[4796]: E1127 11:59:31.928589 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-api" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.928671 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-api" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.928995 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-log" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.929131 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" containerName="nova-api-api" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.930568 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.930854 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.933343 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.933760 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 27 11:59:31 crc kubenswrapper[4796]: I1127 11:59:31.934701 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.074474 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4574506-0fa9-41e7-ace2-0151a43916a5-logs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.074549 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.074760 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.074971 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hwcw\" (UniqueName: \"kubernetes.io/projected/d4574506-0fa9-41e7-ace2-0151a43916a5-kube-api-access-2hwcw\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.075072 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-config-data\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.075189 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176394 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4574506-0fa9-41e7-ace2-0151a43916a5-logs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176476 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176510 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176579 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hwcw\" (UniqueName: \"kubernetes.io/projected/d4574506-0fa9-41e7-ace2-0151a43916a5-kube-api-access-2hwcw\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176622 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-config-data\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.176930 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4574506-0fa9-41e7-ace2-0151a43916a5-logs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.181236 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.181366 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.182023 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.182411 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4574506-0fa9-41e7-ace2-0151a43916a5-config-data\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.203458 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hwcw\" (UniqueName: \"kubernetes.io/projected/d4574506-0fa9-41e7-ace2-0151a43916a5-kube-api-access-2hwcw\") pod \"nova-api-0\" (UID: \"d4574506-0fa9-41e7-ace2-0151a43916a5\") " pod="openstack/nova-api-0" Nov 27 11:59:32 crc kubenswrapper[4796]: I1127 11:59:32.292783 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:32.739661 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 11:59:33 crc kubenswrapper[4796]: W1127 11:59:32.742773 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4574506_0fa9_41e7_ace2_0151a43916a5.slice/crio-9491d3d06fb22735105489a6d79a4a4ba5a3d98943c6d0ed0495de6e3099418b WatchSource:0}: Error finding container 9491d3d06fb22735105489a6d79a4a4ba5a3d98943c6d0ed0495de6e3099418b: Status 404 returned error can't find the container with id 9491d3d06fb22735105489a6d79a4a4ba5a3d98943c6d0ed0495de6e3099418b Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:32.852946 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4574506-0fa9-41e7-ace2-0151a43916a5","Type":"ContainerStarted","Data":"9491d3d06fb22735105489a6d79a4a4ba5a3d98943c6d0ed0495de6e3099418b"} Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:32.857832 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerStarted","Data":"55c2f0680d2d54bb4f7fb83c6c998b02b8141f5e0f2f28fab4d35662bb70c576"} Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:33.584797 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a33aa29-4670-4c44-bbae-428de553ca8f" path="/var/lib/kubelet/pods/4a33aa29-4670-4c44-bbae-428de553ca8f/volumes" Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:33.873824 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerStarted","Data":"14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d"} Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:33.873888 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerStarted","Data":"9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8"} Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:33.876737 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4574506-0fa9-41e7-ace2-0151a43916a5","Type":"ContainerStarted","Data":"a9846b80127bc45d6a6a4479e6b5d122fe9429e4ce08a18e42f0b752ee52edf5"} Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:33.876810 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4574506-0fa9-41e7-ace2-0151a43916a5","Type":"ContainerStarted","Data":"0050520ed00589a40d019e13e19bd4428f455b40ac8d5978ab932cea5cc90b7d"} Nov 27 11:59:33 crc kubenswrapper[4796]: I1127 11:59:33.903970 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.903937119 podStartE2EDuration="2.903937119s" podCreationTimestamp="2025-11-27 11:59:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:33.898753918 +0000 UTC m=+2091.417072836" watchObservedRunningTime="2025-11-27 11:59:33.903937119 +0000 UTC m=+2091.422256077" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.204409 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:43424->10.217.0.201:8775: read: connection reset by peer" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.205339 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:43438->10.217.0.201:8775: read: connection reset by peer" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.703163 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.833255 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-config-data\") pod \"5a5c6984-1064-4944-a335-9cd8fa5c1738\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.833415 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5c6984-1064-4944-a335-9cd8fa5c1738-logs\") pod \"5a5c6984-1064-4944-a335-9cd8fa5c1738\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.833480 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx4rv\" (UniqueName: \"kubernetes.io/projected/5a5c6984-1064-4944-a335-9cd8fa5c1738-kube-api-access-gx4rv\") pod \"5a5c6984-1064-4944-a335-9cd8fa5c1738\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.833640 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-combined-ca-bundle\") pod \"5a5c6984-1064-4944-a335-9cd8fa5c1738\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.834108 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a5c6984-1064-4944-a335-9cd8fa5c1738-logs" (OuterVolumeSpecName: "logs") pod "5a5c6984-1064-4944-a335-9cd8fa5c1738" (UID: "5a5c6984-1064-4944-a335-9cd8fa5c1738"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.834236 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-nova-metadata-tls-certs\") pod \"5a5c6984-1064-4944-a335-9cd8fa5c1738\" (UID: \"5a5c6984-1064-4944-a335-9cd8fa5c1738\") " Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.835116 4796 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5c6984-1064-4944-a335-9cd8fa5c1738-logs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.841456 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a5c6984-1064-4944-a335-9cd8fa5c1738-kube-api-access-gx4rv" (OuterVolumeSpecName: "kube-api-access-gx4rv") pod "5a5c6984-1064-4944-a335-9cd8fa5c1738" (UID: "5a5c6984-1064-4944-a335-9cd8fa5c1738"). InnerVolumeSpecName "kube-api-access-gx4rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.867591 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-config-data" (OuterVolumeSpecName: "config-data") pod "5a5c6984-1064-4944-a335-9cd8fa5c1738" (UID: "5a5c6984-1064-4944-a335-9cd8fa5c1738"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.869245 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a5c6984-1064-4944-a335-9cd8fa5c1738" (UID: "5a5c6984-1064-4944-a335-9cd8fa5c1738"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.888529 4796 generic.go:334] "Generic (PLEG): container finished" podID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerID="83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026" exitCode=0 Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.888646 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.889539 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5a5c6984-1064-4944-a335-9cd8fa5c1738","Type":"ContainerDied","Data":"83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026"} Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.889578 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5a5c6984-1064-4944-a335-9cd8fa5c1738","Type":"ContainerDied","Data":"588d6740ab97c59d37dba741e099bef9274db88c3b3a62ad5cdd2f6f19c225c9"} Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.889599 4796 scope.go:117] "RemoveContainer" containerID="83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.892931 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerStarted","Data":"4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132"} Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.899667 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5a5c6984-1064-4944-a335-9cd8fa5c1738" (UID: "5a5c6984-1064-4944-a335-9cd8fa5c1738"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.924717 4796 scope.go:117] "RemoveContainer" containerID="796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.937418 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.937456 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx4rv\" (UniqueName: \"kubernetes.io/projected/5a5c6984-1064-4944-a335-9cd8fa5c1738-kube-api-access-gx4rv\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.937467 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.937479 4796 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5c6984-1064-4944-a335-9cd8fa5c1738-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.944856 4796 scope.go:117] "RemoveContainer" containerID="83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026" Nov 27 11:59:34 crc kubenswrapper[4796]: E1127 11:59:34.945703 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026\": container with ID starting with 83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026 not found: ID does not exist" containerID="83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.945745 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026"} err="failed to get container status \"83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026\": rpc error: code = NotFound desc = could not find container \"83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026\": container with ID starting with 83ac5d3d0be0d6e8719b216e2d7da4696916154b585a30ca5999d3f29dc60026 not found: ID does not exist" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.945770 4796 scope.go:117] "RemoveContainer" containerID="796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254" Nov 27 11:59:34 crc kubenswrapper[4796]: E1127 11:59:34.946449 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254\": container with ID starting with 796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254 not found: ID does not exist" containerID="796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.946474 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254"} err="failed to get container status \"796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254\": rpc error: code = NotFound desc = could not find container \"796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254\": container with ID starting with 796f42f7b8449168fb14b61f2f00de99146c44c82474d26573c2657b7f0aa254 not found: ID does not exist" Nov 27 11:59:34 crc kubenswrapper[4796]: I1127 11:59:34.952649 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.223911 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.233959 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.248331 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:59:35 crc kubenswrapper[4796]: E1127 11:59:35.248822 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-metadata" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.248847 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-metadata" Nov 27 11:59:35 crc kubenswrapper[4796]: E1127 11:59:35.248871 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-log" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.248880 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-log" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.249120 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-metadata" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.249171 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" containerName="nova-metadata-log" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.250566 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.256721 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.256973 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.262494 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.344352 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.344545 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-config-data\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.344606 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d98b2\" (UniqueName: \"kubernetes.io/projected/12250a93-d356-4097-b50e-57343bc8b2c9-kube-api-access-d98b2\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.344642 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.344676 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12250a93-d356-4097-b50e-57343bc8b2c9-logs\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.445841 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12250a93-d356-4097-b50e-57343bc8b2c9-logs\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.445940 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.445997 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-config-data\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.446049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d98b2\" (UniqueName: \"kubernetes.io/projected/12250a93-d356-4097-b50e-57343bc8b2c9-kube-api-access-d98b2\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.446082 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.447309 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12250a93-d356-4097-b50e-57343bc8b2c9-logs\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.450993 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.451484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.451872 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12250a93-d356-4097-b50e-57343bc8b2c9-config-data\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.480837 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d98b2\" (UniqueName: \"kubernetes.io/projected/12250a93-d356-4097-b50e-57343bc8b2c9-kube-api-access-d98b2\") pod \"nova-metadata-0\" (UID: \"12250a93-d356-4097-b50e-57343bc8b2c9\") " pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.568846 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.580606 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a5c6984-1064-4944-a335-9cd8fa5c1738" path="/var/lib/kubelet/pods/5a5c6984-1064-4944-a335-9cd8fa5c1738/volumes" Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.911084 4796 generic.go:334] "Generic (PLEG): container finished" podID="406db992-3094-4324-a2fb-ed288d3ee288" containerID="df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d" exitCode=0 Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.911255 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406db992-3094-4324-a2fb-ed288d3ee288","Type":"ContainerDied","Data":"df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d"} Nov 27 11:59:35 crc kubenswrapper[4796]: I1127 11:59:35.983834 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:59:36 crc kubenswrapper[4796]: W1127 11:59:36.010926 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12250a93_d356_4097_b50e_57343bc8b2c9.slice/crio-141fc89884697436d2b50b02e15dc57633042123413e6affaddb0b1ce81a9ec8 WatchSource:0}: Error finding container 141fc89884697436d2b50b02e15dc57633042123413e6affaddb0b1ce81a9ec8: Status 404 returned error can't find the container with id 141fc89884697436d2b50b02e15dc57633042123413e6affaddb0b1ce81a9ec8 Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.012352 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.058326 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvbh9\" (UniqueName: \"kubernetes.io/projected/406db992-3094-4324-a2fb-ed288d3ee288-kube-api-access-lvbh9\") pod \"406db992-3094-4324-a2fb-ed288d3ee288\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.058498 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-config-data\") pod \"406db992-3094-4324-a2fb-ed288d3ee288\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.058547 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-combined-ca-bundle\") pod \"406db992-3094-4324-a2fb-ed288d3ee288\" (UID: \"406db992-3094-4324-a2fb-ed288d3ee288\") " Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.069414 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406db992-3094-4324-a2fb-ed288d3ee288-kube-api-access-lvbh9" (OuterVolumeSpecName: "kube-api-access-lvbh9") pod "406db992-3094-4324-a2fb-ed288d3ee288" (UID: "406db992-3094-4324-a2fb-ed288d3ee288"). InnerVolumeSpecName "kube-api-access-lvbh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.093295 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "406db992-3094-4324-a2fb-ed288d3ee288" (UID: "406db992-3094-4324-a2fb-ed288d3ee288"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.120522 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-config-data" (OuterVolumeSpecName: "config-data") pod "406db992-3094-4324-a2fb-ed288d3ee288" (UID: "406db992-3094-4324-a2fb-ed288d3ee288"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.160966 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvbh9\" (UniqueName: \"kubernetes.io/projected/406db992-3094-4324-a2fb-ed288d3ee288-kube-api-access-lvbh9\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.161004 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.161017 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406db992-3094-4324-a2fb-ed288d3ee288-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.923679 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12250a93-d356-4097-b50e-57343bc8b2c9","Type":"ContainerStarted","Data":"1daf3cf88896a1482aeefb09910a0466836de11a47e5f4dd3f68fa1b8b01669c"} Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.924943 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12250a93-d356-4097-b50e-57343bc8b2c9","Type":"ContainerStarted","Data":"3d1ee570e4b883c6e2ceb21ad0f5c05709cf56c3f27b0194223cfee10a437599"} Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.925014 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"12250a93-d356-4097-b50e-57343bc8b2c9","Type":"ContainerStarted","Data":"141fc89884697436d2b50b02e15dc57633042123413e6affaddb0b1ce81a9ec8"} Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.928704 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerStarted","Data":"66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664"} Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.928956 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.930612 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406db992-3094-4324-a2fb-ed288d3ee288","Type":"ContainerDied","Data":"ff741c9701221d4513d8a19142261568a74481c6a1629220fa949352bd6203c6"} Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.930677 4796 scope.go:117] "RemoveContainer" containerID="df28d58f0f9f29005aa2eef850b01b6bce615ef9af9010bf07fe8d7b0d50ea7d" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.930805 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.958136 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.95811829 podStartE2EDuration="1.95811829s" podCreationTimestamp="2025-11-27 11:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:36.947510441 +0000 UTC m=+2094.465829359" watchObservedRunningTime="2025-11-27 11:59:36.95811829 +0000 UTC m=+2094.476437228" Nov 27 11:59:36 crc kubenswrapper[4796]: I1127 11:59:36.988514 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.007831 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.028984 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:59:37 crc kubenswrapper[4796]: E1127 11:59:37.029554 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406db992-3094-4324-a2fb-ed288d3ee288" containerName="nova-scheduler-scheduler" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.029581 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="406db992-3094-4324-a2fb-ed288d3ee288" containerName="nova-scheduler-scheduler" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.029789 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="406db992-3094-4324-a2fb-ed288d3ee288" containerName="nova-scheduler-scheduler" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.030584 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.033876 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.476726937 podStartE2EDuration="7.033858075s" podCreationTimestamp="2025-11-27 11:59:30 +0000 UTC" firstStartedPulling="2025-11-27 11:59:31.897713874 +0000 UTC m=+2089.416032802" lastFinishedPulling="2025-11-27 11:59:36.454845022 +0000 UTC m=+2093.973163940" observedRunningTime="2025-11-27 11:59:37.007141306 +0000 UTC m=+2094.525460224" watchObservedRunningTime="2025-11-27 11:59:37.033858075 +0000 UTC m=+2094.552176993" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.037718 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.045498 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.081942 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47781b1-daeb-4448-b7c5-55005abc5f2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.082023 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b47781b1-daeb-4448-b7c5-55005abc5f2e-config-data\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.082069 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svcw9\" (UniqueName: \"kubernetes.io/projected/b47781b1-daeb-4448-b7c5-55005abc5f2e-kube-api-access-svcw9\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.183463 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47781b1-daeb-4448-b7c5-55005abc5f2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.183801 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b47781b1-daeb-4448-b7c5-55005abc5f2e-config-data\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.183847 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svcw9\" (UniqueName: \"kubernetes.io/projected/b47781b1-daeb-4448-b7c5-55005abc5f2e-kube-api-access-svcw9\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.187461 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b47781b1-daeb-4448-b7c5-55005abc5f2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.188220 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b47781b1-daeb-4448-b7c5-55005abc5f2e-config-data\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.203506 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svcw9\" (UniqueName: \"kubernetes.io/projected/b47781b1-daeb-4448-b7c5-55005abc5f2e-kube-api-access-svcw9\") pod \"nova-scheduler-0\" (UID: \"b47781b1-daeb-4448-b7c5-55005abc5f2e\") " pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.353620 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.583868 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406db992-3094-4324-a2fb-ed288d3ee288" path="/var/lib/kubelet/pods/406db992-3094-4324-a2fb-ed288d3ee288/volumes" Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.778972 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 11:59:37 crc kubenswrapper[4796]: W1127 11:59:37.781257 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb47781b1_daeb_4448_b7c5_55005abc5f2e.slice/crio-fe85330348c44b973bca609c15fcee41167901c281ccbc1de84399afbc7a0747 WatchSource:0}: Error finding container fe85330348c44b973bca609c15fcee41167901c281ccbc1de84399afbc7a0747: Status 404 returned error can't find the container with id fe85330348c44b973bca609c15fcee41167901c281ccbc1de84399afbc7a0747 Nov 27 11:59:37 crc kubenswrapper[4796]: I1127 11:59:37.957858 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b47781b1-daeb-4448-b7c5-55005abc5f2e","Type":"ContainerStarted","Data":"fe85330348c44b973bca609c15fcee41167901c281ccbc1de84399afbc7a0747"} Nov 27 11:59:38 crc kubenswrapper[4796]: I1127 11:59:38.970611 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b47781b1-daeb-4448-b7c5-55005abc5f2e","Type":"ContainerStarted","Data":"ec19ab33c724d91fe7502b9dd5fb3b131266359a13d0219c6906b4c56d159b4d"} Nov 27 11:59:39 crc kubenswrapper[4796]: I1127 11:59:39.001477 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.001454037 podStartE2EDuration="3.001454037s" podCreationTimestamp="2025-11-27 11:59:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:59:38.985625925 +0000 UTC m=+2096.503944843" watchObservedRunningTime="2025-11-27 11:59:39.001454037 +0000 UTC m=+2096.519772955" Nov 27 11:59:40 crc kubenswrapper[4796]: I1127 11:59:40.568991 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 11:59:40 crc kubenswrapper[4796]: I1127 11:59:40.569421 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 11:59:42 crc kubenswrapper[4796]: I1127 11:59:42.293605 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 11:59:42 crc kubenswrapper[4796]: I1127 11:59:42.293866 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 11:59:42 crc kubenswrapper[4796]: I1127 11:59:42.354601 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 11:59:43 crc kubenswrapper[4796]: I1127 11:59:43.304498 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d4574506-0fa9-41e7-ace2-0151a43916a5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:43 crc kubenswrapper[4796]: I1127 11:59:43.304502 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d4574506-0fa9-41e7-ace2-0151a43916a5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:45 crc kubenswrapper[4796]: I1127 11:59:45.589310 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 11:59:45 crc kubenswrapper[4796]: I1127 11:59:45.589738 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 11:59:46 crc kubenswrapper[4796]: I1127 11:59:46.585560 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="12250a93-d356-4097-b50e-57343bc8b2c9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:46 crc kubenswrapper[4796]: I1127 11:59:46.585595 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="12250a93-d356-4097-b50e-57343bc8b2c9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 11:59:47 crc kubenswrapper[4796]: I1127 11:59:47.354591 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 11:59:47 crc kubenswrapper[4796]: I1127 11:59:47.381307 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 11:59:48 crc kubenswrapper[4796]: I1127 11:59:48.087019 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 11:59:52 crc kubenswrapper[4796]: I1127 11:59:52.310578 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 11:59:52 crc kubenswrapper[4796]: I1127 11:59:52.311571 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 11:59:52 crc kubenswrapper[4796]: I1127 11:59:52.322959 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 11:59:52 crc kubenswrapper[4796]: I1127 11:59:52.326485 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 11:59:53 crc kubenswrapper[4796]: I1127 11:59:53.108449 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 11:59:53 crc kubenswrapper[4796]: I1127 11:59:53.114823 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 11:59:55 crc kubenswrapper[4796]: I1127 11:59:55.583666 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 11:59:55 crc kubenswrapper[4796]: I1127 11:59:55.584075 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 11:59:55 crc kubenswrapper[4796]: I1127 11:59:55.591337 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 11:59:55 crc kubenswrapper[4796]: I1127 11:59:55.591819 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.161013 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5"] Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.163552 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.169128 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.169706 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.174888 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5"] Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.246695 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-secret-volume\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.246857 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64npv\" (UniqueName: \"kubernetes.io/projected/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-kube-api-access-64npv\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.247047 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-config-volume\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.348808 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-secret-volume\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.349178 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64npv\" (UniqueName: \"kubernetes.io/projected/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-kube-api-access-64npv\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.349343 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-config-volume\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.350406 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-config-volume\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.358586 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-secret-volume\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.369664 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64npv\" (UniqueName: \"kubernetes.io/projected/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-kube-api-access-64npv\") pod \"collect-profiles-29404080-vh5j5\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:00 crc kubenswrapper[4796]: I1127 12:00:00.551364 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:01 crc kubenswrapper[4796]: I1127 12:00:01.024381 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5"] Nov 27 12:00:01 crc kubenswrapper[4796]: W1127 12:00:01.031468 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4228ea4e_2713_4d86_8922_8e4e5e4fbdbe.slice/crio-6afcae2162348a6b33ee56dfcdf26ee100fd965b54c853d8d6033748550aaa2e WatchSource:0}: Error finding container 6afcae2162348a6b33ee56dfcdf26ee100fd965b54c853d8d6033748550aaa2e: Status 404 returned error can't find the container with id 6afcae2162348a6b33ee56dfcdf26ee100fd965b54c853d8d6033748550aaa2e Nov 27 12:00:01 crc kubenswrapper[4796]: I1127 12:00:01.208929 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" event={"ID":"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe","Type":"ContainerStarted","Data":"c819fb7b4d9217b8e4b5242ed7e2363d0dca8ff8737017affafebe3172cba86e"} Nov 27 12:00:01 crc kubenswrapper[4796]: I1127 12:00:01.208980 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" event={"ID":"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe","Type":"ContainerStarted","Data":"6afcae2162348a6b33ee56dfcdf26ee100fd965b54c853d8d6033748550aaa2e"} Nov 27 12:00:01 crc kubenswrapper[4796]: I1127 12:00:01.231251 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" podStartSLOduration=1.231227836 podStartE2EDuration="1.231227836s" podCreationTimestamp="2025-11-27 12:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:00:01.222213721 +0000 UTC m=+2118.740532639" watchObservedRunningTime="2025-11-27 12:00:01.231227836 +0000 UTC m=+2118.749546754" Nov 27 12:00:01 crc kubenswrapper[4796]: I1127 12:00:01.370609 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 12:00:02 crc kubenswrapper[4796]: I1127 12:00:02.218559 4796 generic.go:334] "Generic (PLEG): container finished" podID="4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" containerID="c819fb7b4d9217b8e4b5242ed7e2363d0dca8ff8737017affafebe3172cba86e" exitCode=0 Nov 27 12:00:02 crc kubenswrapper[4796]: I1127 12:00:02.218626 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" event={"ID":"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe","Type":"ContainerDied","Data":"c819fb7b4d9217b8e4b5242ed7e2363d0dca8ff8737017affafebe3172cba86e"} Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.558134 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.736249 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64npv\" (UniqueName: \"kubernetes.io/projected/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-kube-api-access-64npv\") pod \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.736508 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-config-volume\") pod \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.736734 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-secret-volume\") pod \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\" (UID: \"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe\") " Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.737768 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-config-volume" (OuterVolumeSpecName: "config-volume") pod "4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" (UID: "4228ea4e-2713-4d86-8922-8e4e5e4fbdbe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.744873 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" (UID: "4228ea4e-2713-4d86-8922-8e4e5e4fbdbe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.746756 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-kube-api-access-64npv" (OuterVolumeSpecName: "kube-api-access-64npv") pod "4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" (UID: "4228ea4e-2713-4d86-8922-8e4e5e4fbdbe"). InnerVolumeSpecName "kube-api-access-64npv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.838486 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.838552 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64npv\" (UniqueName: \"kubernetes.io/projected/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-kube-api-access-64npv\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:03 crc kubenswrapper[4796]: I1127 12:00:03.838567 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:04 crc kubenswrapper[4796]: I1127 12:00:04.237481 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" event={"ID":"4228ea4e-2713-4d86-8922-8e4e5e4fbdbe","Type":"ContainerDied","Data":"6afcae2162348a6b33ee56dfcdf26ee100fd965b54c853d8d6033748550aaa2e"} Nov 27 12:00:04 crc kubenswrapper[4796]: I1127 12:00:04.237517 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5" Nov 27 12:00:04 crc kubenswrapper[4796]: I1127 12:00:04.237525 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6afcae2162348a6b33ee56dfcdf26ee100fd965b54c853d8d6033748550aaa2e" Nov 27 12:00:04 crc kubenswrapper[4796]: I1127 12:00:04.293981 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk"] Nov 27 12:00:04 crc kubenswrapper[4796]: I1127 12:00:04.305038 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404035-jj5kk"] Nov 27 12:00:05 crc kubenswrapper[4796]: I1127 12:00:05.584727 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9297600d-7a17-4a87-9ed1-26357ea3df51" path="/var/lib/kubelet/pods/9297600d-7a17-4a87-9ed1-26357ea3df51/volumes" Nov 27 12:00:11 crc kubenswrapper[4796]: I1127 12:00:11.015508 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 12:00:11 crc kubenswrapper[4796]: I1127 12:00:11.825416 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 12:00:15 crc kubenswrapper[4796]: I1127 12:00:15.187235 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="rabbitmq" containerID="cri-o://0cd3cd3cb58ddab12d8a55ce4bcef4301b0141513e2fc0b859404aed61f14527" gracePeriod=604796 Nov 27 12:00:15 crc kubenswrapper[4796]: I1127 12:00:15.912984 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="rabbitmq" containerID="cri-o://f89b18a271e227ede16b294522eb260aa74c47f021952b0da1f556debaf2c649" gracePeriod=604796 Nov 27 12:00:17 crc kubenswrapper[4796]: I1127 12:00:17.189757 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 27 12:00:17 crc kubenswrapper[4796]: I1127 12:00:17.751039 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Nov 27 12:00:20 crc kubenswrapper[4796]: I1127 12:00:20.619653 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6568564dcc-mst5g" podUID="c65b890a-db52-4342-938c-d56538e76b3c" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 27 12:00:21 crc kubenswrapper[4796]: I1127 12:00:21.449774 4796 generic.go:334] "Generic (PLEG): container finished" podID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerID="0cd3cd3cb58ddab12d8a55ce4bcef4301b0141513e2fc0b859404aed61f14527" exitCode=0 Nov 27 12:00:21 crc kubenswrapper[4796]: I1127 12:00:21.449831 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b853d8bd-9978-45a7-ad7c-b3217d4a93b3","Type":"ContainerDied","Data":"0cd3cd3cb58ddab12d8a55ce4bcef4301b0141513e2fc0b859404aed61f14527"} Nov 27 12:00:21 crc kubenswrapper[4796]: I1127 12:00:21.899381 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012592 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-server-conf\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012718 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-pod-info\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012770 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-confd\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012847 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-plugins-conf\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012911 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-erlang-cookie\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012954 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-tls\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.012983 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.013017 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-config-data\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.013072 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-erlang-cookie-secret\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.013139 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq8jg\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-kube-api-access-nq8jg\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.013178 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-plugins\") pod \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\" (UID: \"b853d8bd-9978-45a7-ad7c-b3217d4a93b3\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.013419 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.013681 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.014396 4796 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.014425 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.015189 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.021398 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.026092 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.035142 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.043788 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-pod-info" (OuterVolumeSpecName: "pod-info") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.048534 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-kube-api-access-nq8jg" (OuterVolumeSpecName: "kube-api-access-nq8jg") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "kube-api-access-nq8jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.115783 4796 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.116247 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.116372 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.116469 4796 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.116549 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq8jg\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-kube-api-access-nq8jg\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.116622 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.120845 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-config-data" (OuterVolumeSpecName: "config-data") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.163656 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.171869 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-server-conf" (OuterVolumeSpecName: "server-conf") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.213634 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b853d8bd-9978-45a7-ad7c-b3217d4a93b3" (UID: "b853d8bd-9978-45a7-ad7c-b3217d4a93b3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.219120 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.219173 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.219187 4796 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-server-conf\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.219197 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b853d8bd-9978-45a7-ad7c-b3217d4a93b3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.459810 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b853d8bd-9978-45a7-ad7c-b3217d4a93b3","Type":"ContainerDied","Data":"a9001c375a0b55a1ac3a65d849e89280d854a1d87ca9594984d173bfd1a52585"} Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.459855 4796 scope.go:117] "RemoveContainer" containerID="0cd3cd3cb58ddab12d8a55ce4bcef4301b0141513e2fc0b859404aed61f14527" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.459888 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.464091 4796 generic.go:334] "Generic (PLEG): container finished" podID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerID="f89b18a271e227ede16b294522eb260aa74c47f021952b0da1f556debaf2c649" exitCode=0 Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.464149 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e9cd74f-d96d-4be4-b448-01d027b34689","Type":"ContainerDied","Data":"f89b18a271e227ede16b294522eb260aa74c47f021952b0da1f556debaf2c649"} Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.464184 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5e9cd74f-d96d-4be4-b448-01d027b34689","Type":"ContainerDied","Data":"2d1a4349c8fa3954bd9222adf0dd39a1ffea03fc072dd421a2afaaba355dc2f6"} Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.464200 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d1a4349c8fa3954bd9222adf0dd39a1ffea03fc072dd421a2afaaba355dc2f6" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.537556 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.550232 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.552070 4796 scope.go:117] "RemoveContainer" containerID="c189a5bb7656fc2b2a2ccbdbd7bc017c1a6d0494b20cf5077e9aa8b7abe84c46" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.561432 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.628708 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 12:00:22 crc kubenswrapper[4796]: E1127 12:00:22.629235 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="rabbitmq" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629249 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="rabbitmq" Nov 27 12:00:22 crc kubenswrapper[4796]: E1127 12:00:22.629278 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="rabbitmq" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629284 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="rabbitmq" Nov 27 12:00:22 crc kubenswrapper[4796]: E1127 12:00:22.629294 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="setup-container" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629301 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="setup-container" Nov 27 12:00:22 crc kubenswrapper[4796]: E1127 12:00:22.629337 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" containerName="collect-profiles" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629343 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" containerName="collect-profiles" Nov 27 12:00:22 crc kubenswrapper[4796]: E1127 12:00:22.629356 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="setup-container" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629361 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="setup-container" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629519 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" containerName="rabbitmq" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629532 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" containerName="rabbitmq" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.629549 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" containerName="collect-profiles" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.630516 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.636975 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.636978 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637373 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637387 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-confd\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637549 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e9cd74f-d96d-4be4-b448-01d027b34689-erlang-cookie-secret\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637628 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e9cd74f-d96d-4be4-b448-01d027b34689-pod-info\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637662 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjhvl\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-kube-api-access-cjhvl\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637708 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-server-conf\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637751 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-plugins\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637801 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637826 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-erlang-cookie\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637856 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-plugins-conf\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637885 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-config-data\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.637904 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-tls\") pod \"5e9cd74f-d96d-4be4-b448-01d027b34689\" (UID: \"5e9cd74f-d96d-4be4-b448-01d027b34689\") " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.638340 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.638593 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.638776 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.639515 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.639832 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.640909 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.641135 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h228j" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.642519 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.657220 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.659912 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.661915 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.662130 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-kube-api-access-cjhvl" (OuterVolumeSpecName: "kube-api-access-cjhvl") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "kube-api-access-cjhvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.662506 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5e9cd74f-d96d-4be4-b448-01d027b34689-pod-info" (OuterVolumeSpecName: "pod-info") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.700035 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e9cd74f-d96d-4be4-b448-01d027b34689-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.728948 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-config-data" (OuterVolumeSpecName: "config-data") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.740092 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-config-data\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.740327 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.742972 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.743129 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a340b9d9-d859-4b22-a55a-e955b7951806-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.743702 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcl5r\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-kube-api-access-lcl5r\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.743796 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.743874 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744022 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a340b9d9-d859-4b22-a55a-e955b7951806-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744139 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744282 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744438 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744595 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744670 4796 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e9cd74f-d96d-4be4-b448-01d027b34689-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744729 4796 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e9cd74f-d96d-4be4-b448-01d027b34689-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744888 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjhvl\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-kube-api-access-cjhvl\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.744970 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.745040 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.745092 4796 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.745159 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.770836 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-server-conf" (OuterVolumeSpecName: "server-conf") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.773979 4796 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.842121 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5e9cd74f-d96d-4be4-b448-01d027b34689" (UID: "5e9cd74f-d96d-4be4-b448-01d027b34689"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846314 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-config-data\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846380 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846415 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846459 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a340b9d9-d859-4b22-a55a-e955b7951806-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846493 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcl5r\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-kube-api-access-lcl5r\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846518 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846539 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846580 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a340b9d9-d859-4b22-a55a-e955b7951806-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846620 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846637 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846667 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846724 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846799 4796 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e9cd74f-d96d-4be4-b448-01d027b34689-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846815 4796 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e9cd74f-d96d-4be4-b448-01d027b34689-server-conf\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.846827 4796 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.847438 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-config-data\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.847548 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.847952 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.848421 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.849237 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a340b9d9-d859-4b22-a55a-e955b7951806-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.851915 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a340b9d9-d859-4b22-a55a-e955b7951806-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.853200 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.853613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a340b9d9-d859-4b22-a55a-e955b7951806-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.855101 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.869865 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcl5r\" (UniqueName: \"kubernetes.io/projected/a340b9d9-d859-4b22-a55a-e955b7951806-kube-api-access-lcl5r\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:22 crc kubenswrapper[4796]: I1127 12:00:22.896967 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"a340b9d9-d859-4b22-a55a-e955b7951806\") " pod="openstack/rabbitmq-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.098592 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.475953 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.520513 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.546340 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.556663 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.558152 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.560730 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.560993 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.561221 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.562428 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.563240 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.563577 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.563704 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-b955p" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.586255 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e9cd74f-d96d-4be4-b448-01d027b34689" path="/var/lib/kubelet/pods/5e9cd74f-d96d-4be4-b448-01d027b34689/volumes" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.587375 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b853d8bd-9978-45a7-ad7c-b3217d4a93b3" path="/var/lib/kubelet/pods/b853d8bd-9978-45a7-ad7c-b3217d4a93b3/volumes" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.588321 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.611196 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659396 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659478 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659501 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659525 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659550 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659571 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659604 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659632 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659671 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659751 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.659783 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8l7b\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-kube-api-access-t8l7b\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762060 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762158 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762214 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762238 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762253 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762725 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762739 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762764 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762799 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.762850 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.763174 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.763254 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8l7b\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-kube-api-access-t8l7b\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.763430 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.763850 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.763988 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.764239 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.764887 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.766503 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.767160 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.770191 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.771067 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.791342 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8l7b\" (UniqueName: \"kubernetes.io/projected/9f4f2a8e-1ba5-4af3-8154-02a911ca9d14-kube-api-access-t8l7b\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.804456 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:23 crc kubenswrapper[4796]: I1127 12:00:23.880743 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:00:24 crc kubenswrapper[4796]: W1127 12:00:24.306835 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f4f2a8e_1ba5_4af3_8154_02a911ca9d14.slice/crio-e6fecd504045315436cba60bdd22fd595dadbd91c047cc9fedc5cbf21aa5e06b WatchSource:0}: Error finding container e6fecd504045315436cba60bdd22fd595dadbd91c047cc9fedc5cbf21aa5e06b: Status 404 returned error can't find the container with id e6fecd504045315436cba60bdd22fd595dadbd91c047cc9fedc5cbf21aa5e06b Nov 27 12:00:24 crc kubenswrapper[4796]: I1127 12:00:24.309633 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 12:00:24 crc kubenswrapper[4796]: I1127 12:00:24.488357 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a340b9d9-d859-4b22-a55a-e955b7951806","Type":"ContainerStarted","Data":"d52acbba42766b90342d60a646979db582592dff080aa9b840108c05ff2fabe9"} Nov 27 12:00:24 crc kubenswrapper[4796]: I1127 12:00:24.490229 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14","Type":"ContainerStarted","Data":"e6fecd504045315436cba60bdd22fd595dadbd91c047cc9fedc5cbf21aa5e06b"} Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.098116 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8zpz2"] Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.100513 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.102298 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.120310 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8zpz2"] Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193095 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-config\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193456 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193527 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193561 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd9dj\" (UniqueName: \"kubernetes.io/projected/775f764f-9c60-4116-b261-b18fefb303fb-kube-api-access-wd9dj\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193592 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.193632 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295561 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295621 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295723 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295743 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-config\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295786 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295824 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.295846 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd9dj\" (UniqueName: \"kubernetes.io/projected/775f764f-9c60-4116-b261-b18fefb303fb-kube-api-access-wd9dj\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.296683 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.296872 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-config\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.296878 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.296879 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.296992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.297352 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.472405 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd9dj\" (UniqueName: \"kubernetes.io/projected/775f764f-9c60-4116-b261-b18fefb303fb-kube-api-access-wd9dj\") pod \"dnsmasq-dns-5b75489c6f-8zpz2\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.501372 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a340b9d9-d859-4b22-a55a-e955b7951806","Type":"ContainerStarted","Data":"a244eca12e2bd79bc70acefe63618f3d4633bf5c6ac89097bb99f4ad3e27498d"} Nov 27 12:00:25 crc kubenswrapper[4796]: I1127 12:00:25.738124 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:26 crc kubenswrapper[4796]: I1127 12:00:26.181058 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8zpz2"] Nov 27 12:00:26 crc kubenswrapper[4796]: I1127 12:00:26.512286 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14","Type":"ContainerStarted","Data":"b62ceb46c8b043b4d28a0922731936c6a996bc7b0b03a034872864b272f637b2"} Nov 27 12:00:26 crc kubenswrapper[4796]: I1127 12:00:26.515970 4796 generic.go:334] "Generic (PLEG): container finished" podID="775f764f-9c60-4116-b261-b18fefb303fb" containerID="a9bd8410369b1e739382d5971fb4d38d43f143cba6fab7486cd3af9bdc594241" exitCode=0 Nov 27 12:00:26 crc kubenswrapper[4796]: I1127 12:00:26.516038 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" event={"ID":"775f764f-9c60-4116-b261-b18fefb303fb","Type":"ContainerDied","Data":"a9bd8410369b1e739382d5971fb4d38d43f143cba6fab7486cd3af9bdc594241"} Nov 27 12:00:26 crc kubenswrapper[4796]: I1127 12:00:26.516099 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" event={"ID":"775f764f-9c60-4116-b261-b18fefb303fb","Type":"ContainerStarted","Data":"61b649aee58b599870511437c8ac8bd5cc95c28172d987ee17fcd2d0814ca5a6"} Nov 27 12:00:27 crc kubenswrapper[4796]: I1127 12:00:27.530698 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" event={"ID":"775f764f-9c60-4116-b261-b18fefb303fb","Type":"ContainerStarted","Data":"a0093b08d88970f516f5433ff9c8f9b1f92819e6407df2ee784f7041986b2c6b"} Nov 27 12:00:27 crc kubenswrapper[4796]: I1127 12:00:27.531302 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:27 crc kubenswrapper[4796]: I1127 12:00:27.567718 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" podStartSLOduration=2.567696717 podStartE2EDuration="2.567696717s" podCreationTimestamp="2025-11-27 12:00:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:00:27.558023174 +0000 UTC m=+2145.076342112" watchObservedRunningTime="2025-11-27 12:00:27.567696717 +0000 UTC m=+2145.086015645" Nov 27 12:00:35 crc kubenswrapper[4796]: I1127 12:00:35.739438 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:35 crc kubenswrapper[4796]: I1127 12:00:35.823070 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-qfbbw"] Nov 27 12:00:35 crc kubenswrapper[4796]: I1127 12:00:35.823369 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerName="dnsmasq-dns" containerID="cri-o://bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17" gracePeriod=10 Nov 27 12:00:35 crc kubenswrapper[4796]: I1127 12:00:35.968972 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-kdrpj"] Nov 27 12:00:35 crc kubenswrapper[4796]: I1127 12:00:35.973565 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.006178 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-kdrpj"] Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.107824 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.108054 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.108112 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.108296 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsp6w\" (UniqueName: \"kubernetes.io/projected/7dfd0463-f427-4277-b02d-444ef9bd8565-kube-api-access-vsp6w\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.108410 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.108487 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-config\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.108509 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.210428 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-config\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.210754 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.210822 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.210883 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.210912 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.211008 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsp6w\" (UniqueName: \"kubernetes.io/projected/7dfd0463-f427-4277-b02d-444ef9bd8565-kube-api-access-vsp6w\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.211081 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.211463 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-config\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.212017 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.212020 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.212209 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.212636 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.212814 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfd0463-f427-4277-b02d-444ef9bd8565-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.237226 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsp6w\" (UniqueName: \"kubernetes.io/projected/7dfd0463-f427-4277-b02d-444ef9bd8565-kube-api-access-vsp6w\") pod \"dnsmasq-dns-5d75f767dc-kdrpj\" (UID: \"7dfd0463-f427-4277-b02d-444ef9bd8565\") " pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.295552 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.394194 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.528240 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-swift-storage-0\") pod \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.528649 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xshmn\" (UniqueName: \"kubernetes.io/projected/bcb72c79-d085-44f4-96c8-850aaeeb89f4-kube-api-access-xshmn\") pod \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.528678 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-config\") pod \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.528746 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-svc\") pod \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.528763 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-nb\") pod \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.528791 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-sb\") pod \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\" (UID: \"bcb72c79-d085-44f4-96c8-850aaeeb89f4\") " Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.545744 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcb72c79-d085-44f4-96c8-850aaeeb89f4-kube-api-access-xshmn" (OuterVolumeSpecName: "kube-api-access-xshmn") pod "bcb72c79-d085-44f4-96c8-850aaeeb89f4" (UID: "bcb72c79-d085-44f4-96c8-850aaeeb89f4"). InnerVolumeSpecName "kube-api-access-xshmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.585465 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bcb72c79-d085-44f4-96c8-850aaeeb89f4" (UID: "bcb72c79-d085-44f4-96c8-850aaeeb89f4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.589860 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bcb72c79-d085-44f4-96c8-850aaeeb89f4" (UID: "bcb72c79-d085-44f4-96c8-850aaeeb89f4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.600288 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-config" (OuterVolumeSpecName: "config") pod "bcb72c79-d085-44f4-96c8-850aaeeb89f4" (UID: "bcb72c79-d085-44f4-96c8-850aaeeb89f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.613543 4796 generic.go:334] "Generic (PLEG): container finished" podID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerID="bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17" exitCode=0 Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.613585 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" event={"ID":"bcb72c79-d085-44f4-96c8-850aaeeb89f4","Type":"ContainerDied","Data":"bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17"} Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.613610 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" event={"ID":"bcb72c79-d085-44f4-96c8-850aaeeb89f4","Type":"ContainerDied","Data":"968e866e01e2cdbd68146c79e6eedfaaef22fad96d2b728d25f2e10150b5fbfa"} Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.613629 4796 scope.go:117] "RemoveContainer" containerID="bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.613752 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-qfbbw" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.613995 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bcb72c79-d085-44f4-96c8-850aaeeb89f4" (UID: "bcb72c79-d085-44f4-96c8-850aaeeb89f4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.617937 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcb72c79-d085-44f4-96c8-850aaeeb89f4" (UID: "bcb72c79-d085-44f4-96c8-850aaeeb89f4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.631324 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xshmn\" (UniqueName: \"kubernetes.io/projected/bcb72c79-d085-44f4-96c8-850aaeeb89f4-kube-api-access-xshmn\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.631363 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.631372 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.631380 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.631389 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.631397 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcb72c79-d085-44f4-96c8-850aaeeb89f4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.642379 4796 scope.go:117] "RemoveContainer" containerID="9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.661621 4796 scope.go:117] "RemoveContainer" containerID="bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17" Nov 27 12:00:36 crc kubenswrapper[4796]: E1127 12:00:36.662106 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17\": container with ID starting with bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17 not found: ID does not exist" containerID="bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.662158 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17"} err="failed to get container status \"bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17\": rpc error: code = NotFound desc = could not find container \"bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17\": container with ID starting with bf9217fcc8443ded37d44498fa2be44bc2905cde7d3d61ecc622b547bd23db17 not found: ID does not exist" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.662182 4796 scope.go:117] "RemoveContainer" containerID="9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417" Nov 27 12:00:36 crc kubenswrapper[4796]: E1127 12:00:36.662861 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417\": container with ID starting with 9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417 not found: ID does not exist" containerID="9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.662893 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417"} err="failed to get container status \"9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417\": rpc error: code = NotFound desc = could not find container \"9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417\": container with ID starting with 9a3bb05255da8df5f577d645cb6358f40430d97dce70c02e5c5ce995f9352417 not found: ID does not exist" Nov 27 12:00:36 crc kubenswrapper[4796]: I1127 12:00:36.739906 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-kdrpj"] Nov 27 12:00:36 crc kubenswrapper[4796]: W1127 12:00:36.743934 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dfd0463_f427_4277_b02d_444ef9bd8565.slice/crio-c47327a364938e9e22a14cbcf1a9e0c8af4019415d1445ac8fc501f113f20436 WatchSource:0}: Error finding container c47327a364938e9e22a14cbcf1a9e0c8af4019415d1445ac8fc501f113f20436: Status 404 returned error can't find the container with id c47327a364938e9e22a14cbcf1a9e0c8af4019415d1445ac8fc501f113f20436 Nov 27 12:00:37 crc kubenswrapper[4796]: I1127 12:00:37.082048 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-qfbbw"] Nov 27 12:00:37 crc kubenswrapper[4796]: I1127 12:00:37.096363 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-qfbbw"] Nov 27 12:00:37 crc kubenswrapper[4796]: I1127 12:00:37.589975 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" path="/var/lib/kubelet/pods/bcb72c79-d085-44f4-96c8-850aaeeb89f4/volumes" Nov 27 12:00:37 crc kubenswrapper[4796]: I1127 12:00:37.628536 4796 generic.go:334] "Generic (PLEG): container finished" podID="7dfd0463-f427-4277-b02d-444ef9bd8565" containerID="e128ddcd13ea86b2fabfae4fd960624b4bac0d6e54d058a908d9895eb30c146b" exitCode=0 Nov 27 12:00:37 crc kubenswrapper[4796]: I1127 12:00:37.628900 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" event={"ID":"7dfd0463-f427-4277-b02d-444ef9bd8565","Type":"ContainerDied","Data":"e128ddcd13ea86b2fabfae4fd960624b4bac0d6e54d058a908d9895eb30c146b"} Nov 27 12:00:37 crc kubenswrapper[4796]: I1127 12:00:37.629058 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" event={"ID":"7dfd0463-f427-4277-b02d-444ef9bd8565","Type":"ContainerStarted","Data":"c47327a364938e9e22a14cbcf1a9e0c8af4019415d1445ac8fc501f113f20436"} Nov 27 12:00:38 crc kubenswrapper[4796]: I1127 12:00:38.645430 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" event={"ID":"7dfd0463-f427-4277-b02d-444ef9bd8565","Type":"ContainerStarted","Data":"9d232f82820420f2db91412388cd81c83f84522da64cac24c90a87897b9adbca"} Nov 27 12:00:38 crc kubenswrapper[4796]: I1127 12:00:38.645960 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:38 crc kubenswrapper[4796]: I1127 12:00:38.671434 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" podStartSLOduration=3.671410479 podStartE2EDuration="3.671410479s" podCreationTimestamp="2025-11-27 12:00:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:00:38.666607438 +0000 UTC m=+2156.184926396" watchObservedRunningTime="2025-11-27 12:00:38.671410479 +0000 UTC m=+2156.189729417" Nov 27 12:00:46 crc kubenswrapper[4796]: I1127 12:00:46.297466 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d75f767dc-kdrpj" Nov 27 12:00:46 crc kubenswrapper[4796]: I1127 12:00:46.385018 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8zpz2"] Nov 27 12:00:46 crc kubenswrapper[4796]: I1127 12:00:46.385333 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" podUID="775f764f-9c60-4116-b261-b18fefb303fb" containerName="dnsmasq-dns" containerID="cri-o://a0093b08d88970f516f5433ff9c8f9b1f92819e6407df2ee784f7041986b2c6b" gracePeriod=10 Nov 27 12:00:46 crc kubenswrapper[4796]: I1127 12:00:46.749166 4796 generic.go:334] "Generic (PLEG): container finished" podID="775f764f-9c60-4116-b261-b18fefb303fb" containerID="a0093b08d88970f516f5433ff9c8f9b1f92819e6407df2ee784f7041986b2c6b" exitCode=0 Nov 27 12:00:46 crc kubenswrapper[4796]: I1127 12:00:46.749506 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" event={"ID":"775f764f-9c60-4116-b261-b18fefb303fb","Type":"ContainerDied","Data":"a0093b08d88970f516f5433ff9c8f9b1f92819e6407df2ee784f7041986b2c6b"} Nov 27 12:00:46 crc kubenswrapper[4796]: I1127 12:00:46.918238 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.091471 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-sb\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.091995 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-swift-storage-0\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.092074 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-openstack-edpm-ipam\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.092103 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-svc\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.092130 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-config\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.092153 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd9dj\" (UniqueName: \"kubernetes.io/projected/775f764f-9c60-4116-b261-b18fefb303fb-kube-api-access-wd9dj\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.092177 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-nb\") pod \"775f764f-9c60-4116-b261-b18fefb303fb\" (UID: \"775f764f-9c60-4116-b261-b18fefb303fb\") " Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.099158 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/775f764f-9c60-4116-b261-b18fefb303fb-kube-api-access-wd9dj" (OuterVolumeSpecName: "kube-api-access-wd9dj") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "kube-api-access-wd9dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.145878 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-config" (OuterVolumeSpecName: "config") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.147219 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.149953 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.153797 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.156213 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.167622 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "775f764f-9c60-4116-b261-b18fefb303fb" (UID: "775f764f-9c60-4116-b261-b18fefb303fb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194623 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194657 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd9dj\" (UniqueName: \"kubernetes.io/projected/775f764f-9c60-4116-b261-b18fefb303fb-kube-api-access-wd9dj\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194670 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194679 4796 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194690 4796 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194701 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.194709 4796 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/775f764f-9c60-4116-b261-b18fefb303fb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.761421 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" event={"ID":"775f764f-9c60-4116-b261-b18fefb303fb","Type":"ContainerDied","Data":"61b649aee58b599870511437c8ac8bd5cc95c28172d987ee17fcd2d0814ca5a6"} Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.761516 4796 scope.go:117] "RemoveContainer" containerID="a0093b08d88970f516f5433ff9c8f9b1f92819e6407df2ee784f7041986b2c6b" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.761520 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8zpz2" Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.790097 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8zpz2"] Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.806116 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8zpz2"] Nov 27 12:00:47 crc kubenswrapper[4796]: I1127 12:00:47.813020 4796 scope.go:117] "RemoveContainer" containerID="a9bd8410369b1e739382d5971fb4d38d43f143cba6fab7486cd3af9bdc594241" Nov 27 12:00:49 crc kubenswrapper[4796]: I1127 12:00:49.562101 4796 scope.go:117] "RemoveContainer" containerID="299c7052f7d6546f9849981bf98b33114f042c5ea14a12fc07965930b9c95de6" Nov 27 12:00:49 crc kubenswrapper[4796]: I1127 12:00:49.596745 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="775f764f-9c60-4116-b261-b18fefb303fb" path="/var/lib/kubelet/pods/775f764f-9c60-4116-b261-b18fefb303fb/volumes" Nov 27 12:00:57 crc kubenswrapper[4796]: I1127 12:00:57.858932 4796 generic.go:334] "Generic (PLEG): container finished" podID="a340b9d9-d859-4b22-a55a-e955b7951806" containerID="a244eca12e2bd79bc70acefe63618f3d4633bf5c6ac89097bb99f4ad3e27498d" exitCode=0 Nov 27 12:00:57 crc kubenswrapper[4796]: I1127 12:00:57.859015 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a340b9d9-d859-4b22-a55a-e955b7951806","Type":"ContainerDied","Data":"a244eca12e2bd79bc70acefe63618f3d4633bf5c6ac89097bb99f4ad3e27498d"} Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.604970 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb"] Nov 27 12:00:58 crc kubenswrapper[4796]: E1127 12:00:58.605640 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerName="dnsmasq-dns" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.605658 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerName="dnsmasq-dns" Nov 27 12:00:58 crc kubenswrapper[4796]: E1127 12:00:58.605676 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerName="init" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.605685 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerName="init" Nov 27 12:00:58 crc kubenswrapper[4796]: E1127 12:00:58.605698 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775f764f-9c60-4116-b261-b18fefb303fb" containerName="init" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.605704 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="775f764f-9c60-4116-b261-b18fefb303fb" containerName="init" Nov 27 12:00:58 crc kubenswrapper[4796]: E1127 12:00:58.605723 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775f764f-9c60-4116-b261-b18fefb303fb" containerName="dnsmasq-dns" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.605729 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="775f764f-9c60-4116-b261-b18fefb303fb" containerName="dnsmasq-dns" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.605924 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="775f764f-9c60-4116-b261-b18fefb303fb" containerName="dnsmasq-dns" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.605960 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcb72c79-d085-44f4-96c8-850aaeeb89f4" containerName="dnsmasq-dns" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.606555 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.608763 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.608999 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.612668 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.612672 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.642351 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb"] Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.717767 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.718119 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.718177 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwkm2\" (UniqueName: \"kubernetes.io/projected/81f86614-5247-4cb9-ad8c-688c70f862a6-kube-api-access-xwkm2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.718362 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.820049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwkm2\" (UniqueName: \"kubernetes.io/projected/81f86614-5247-4cb9-ad8c-688c70f862a6-kube-api-access-xwkm2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.820160 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.820240 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.820332 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.833885 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.833918 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.834007 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.837069 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwkm2\" (UniqueName: \"kubernetes.io/projected/81f86614-5247-4cb9-ad8c-688c70f862a6-kube-api-access-xwkm2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.872629 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a340b9d9-d859-4b22-a55a-e955b7951806","Type":"ContainerStarted","Data":"58e7280ece3a69b48e1882ae1147d50f3ff0d56861a4defcf4ca2a839acdcd3c"} Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.872854 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.874589 4796 generic.go:334] "Generic (PLEG): container finished" podID="9f4f2a8e-1ba5-4af3-8154-02a911ca9d14" containerID="b62ceb46c8b043b4d28a0922731936c6a996bc7b0b03a034872864b272f637b2" exitCode=0 Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.874693 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14","Type":"ContainerDied","Data":"b62ceb46c8b043b4d28a0922731936c6a996bc7b0b03a034872864b272f637b2"} Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.904637 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.90461918 podStartE2EDuration="36.90461918s" podCreationTimestamp="2025-11-27 12:00:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:00:58.893720653 +0000 UTC m=+2176.412039621" watchObservedRunningTime="2025-11-27 12:00:58.90461918 +0000 UTC m=+2176.422938098" Nov 27 12:00:58 crc kubenswrapper[4796]: I1127 12:00:58.932660 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:00:59 crc kubenswrapper[4796]: W1127 12:00:59.577630 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81f86614_5247_4cb9_ad8c_688c70f862a6.slice/crio-9e82d6f08f61a85b86c6bbb08dbb338ba48e609e645c8b630957807c4cc97815 WatchSource:0}: Error finding container 9e82d6f08f61a85b86c6bbb08dbb338ba48e609e645c8b630957807c4cc97815: Status 404 returned error can't find the container with id 9e82d6f08f61a85b86c6bbb08dbb338ba48e609e645c8b630957807c4cc97815 Nov 27 12:00:59 crc kubenswrapper[4796]: I1127 12:00:59.580436 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:00:59 crc kubenswrapper[4796]: I1127 12:00:59.582554 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb"] Nov 27 12:00:59 crc kubenswrapper[4796]: I1127 12:00:59.884178 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" event={"ID":"81f86614-5247-4cb9-ad8c-688c70f862a6","Type":"ContainerStarted","Data":"9e82d6f08f61a85b86c6bbb08dbb338ba48e609e645c8b630957807c4cc97815"} Nov 27 12:00:59 crc kubenswrapper[4796]: I1127 12:00:59.886279 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4f2a8e-1ba5-4af3-8154-02a911ca9d14","Type":"ContainerStarted","Data":"911d46e8261bff8037b7b8cd411891dbd7b6b2b3ac3a1f33cdce69664043bf70"} Nov 27 12:00:59 crc kubenswrapper[4796]: I1127 12:00:59.907000 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.906979553 podStartE2EDuration="36.906979553s" podCreationTimestamp="2025-11-27 12:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:00:59.906035016 +0000 UTC m=+2177.424353974" watchObservedRunningTime="2025-11-27 12:00:59.906979553 +0000 UTC m=+2177.425298471" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.140650 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29404081-jlkpb"] Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.141772 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.158003 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29404081-jlkpb"] Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.245164 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-config-data\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.245223 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-combined-ca-bundle\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.245405 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5vd5\" (UniqueName: \"kubernetes.io/projected/48aeb330-d74c-4814-9c4d-41e40436e8a3-kube-api-access-f5vd5\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.245521 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-fernet-keys\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.347785 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-fernet-keys\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.347873 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-config-data\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.347921 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-combined-ca-bundle\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.348111 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5vd5\" (UniqueName: \"kubernetes.io/projected/48aeb330-d74c-4814-9c4d-41e40436e8a3-kube-api-access-f5vd5\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.355252 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-fernet-keys\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.355305 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-config-data\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.356150 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-combined-ca-bundle\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.371236 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5vd5\" (UniqueName: \"kubernetes.io/projected/48aeb330-d74c-4814-9c4d-41e40436e8a3-kube-api-access-f5vd5\") pod \"keystone-cron-29404081-jlkpb\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:00 crc kubenswrapper[4796]: I1127 12:01:00.532985 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:01 crc kubenswrapper[4796]: I1127 12:01:01.016677 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29404081-jlkpb"] Nov 27 12:01:01 crc kubenswrapper[4796]: I1127 12:01:01.881827 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:01:01 crc kubenswrapper[4796]: I1127 12:01:01.891616 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:01:01 crc kubenswrapper[4796]: I1127 12:01:01.921683 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404081-jlkpb" event={"ID":"48aeb330-d74c-4814-9c4d-41e40436e8a3","Type":"ContainerStarted","Data":"ff511b407ffc83f7847b4a97a4f23ec849eadec807818c480f04cecacc49e219"} Nov 27 12:01:01 crc kubenswrapper[4796]: I1127 12:01:01.921747 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404081-jlkpb" event={"ID":"48aeb330-d74c-4814-9c4d-41e40436e8a3","Type":"ContainerStarted","Data":"9812323c6539804415e490c8c417703f0d2dbe9304b3781cc1ea2fd6e9b07ac5"} Nov 27 12:01:01 crc kubenswrapper[4796]: I1127 12:01:01.954051 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29404081-jlkpb" podStartSLOduration=1.954031099 podStartE2EDuration="1.954031099s" podCreationTimestamp="2025-11-27 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:01:01.937520799 +0000 UTC m=+2179.455839717" watchObservedRunningTime="2025-11-27 12:01:01.954031099 +0000 UTC m=+2179.472350017" Nov 27 12:01:03 crc kubenswrapper[4796]: I1127 12:01:03.881159 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:01:03 crc kubenswrapper[4796]: I1127 12:01:03.942855 4796 generic.go:334] "Generic (PLEG): container finished" podID="48aeb330-d74c-4814-9c4d-41e40436e8a3" containerID="ff511b407ffc83f7847b4a97a4f23ec849eadec807818c480f04cecacc49e219" exitCode=0 Nov 27 12:01:03 crc kubenswrapper[4796]: I1127 12:01:03.942904 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404081-jlkpb" event={"ID":"48aeb330-d74c-4814-9c4d-41e40436e8a3","Type":"ContainerDied","Data":"ff511b407ffc83f7847b4a97a4f23ec849eadec807818c480f04cecacc49e219"} Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.446594 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.570753 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5vd5\" (UniqueName: \"kubernetes.io/projected/48aeb330-d74c-4814-9c4d-41e40436e8a3-kube-api-access-f5vd5\") pod \"48aeb330-d74c-4814-9c4d-41e40436e8a3\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.570902 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-combined-ca-bundle\") pod \"48aeb330-d74c-4814-9c4d-41e40436e8a3\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.570968 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-fernet-keys\") pod \"48aeb330-d74c-4814-9c4d-41e40436e8a3\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.571006 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-config-data\") pod \"48aeb330-d74c-4814-9c4d-41e40436e8a3\" (UID: \"48aeb330-d74c-4814-9c4d-41e40436e8a3\") " Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.574696 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "48aeb330-d74c-4814-9c4d-41e40436e8a3" (UID: "48aeb330-d74c-4814-9c4d-41e40436e8a3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.589238 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48aeb330-d74c-4814-9c4d-41e40436e8a3-kube-api-access-f5vd5" (OuterVolumeSpecName: "kube-api-access-f5vd5") pod "48aeb330-d74c-4814-9c4d-41e40436e8a3" (UID: "48aeb330-d74c-4814-9c4d-41e40436e8a3"). InnerVolumeSpecName "kube-api-access-f5vd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.625065 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48aeb330-d74c-4814-9c4d-41e40436e8a3" (UID: "48aeb330-d74c-4814-9c4d-41e40436e8a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.642160 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-config-data" (OuterVolumeSpecName: "config-data") pod "48aeb330-d74c-4814-9c4d-41e40436e8a3" (UID: "48aeb330-d74c-4814-9c4d-41e40436e8a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.675891 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.675926 4796 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.676023 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48aeb330-d74c-4814-9c4d-41e40436e8a3-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:09 crc kubenswrapper[4796]: I1127 12:01:09.676043 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5vd5\" (UniqueName: \"kubernetes.io/projected/48aeb330-d74c-4814-9c4d-41e40436e8a3-kube-api-access-f5vd5\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:10 crc kubenswrapper[4796]: I1127 12:01:10.011197 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404081-jlkpb" event={"ID":"48aeb330-d74c-4814-9c4d-41e40436e8a3","Type":"ContainerDied","Data":"9812323c6539804415e490c8c417703f0d2dbe9304b3781cc1ea2fd6e9b07ac5"} Nov 27 12:01:10 crc kubenswrapper[4796]: I1127 12:01:10.011260 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404081-jlkpb" Nov 27 12:01:10 crc kubenswrapper[4796]: I1127 12:01:10.011297 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9812323c6539804415e490c8c417703f0d2dbe9304b3781cc1ea2fd6e9b07ac5" Nov 27 12:01:10 crc kubenswrapper[4796]: I1127 12:01:10.013121 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" event={"ID":"81f86614-5247-4cb9-ad8c-688c70f862a6","Type":"ContainerStarted","Data":"50725dc923c3e968338af0942bc5a8a3a49ffd705db4d1d6c58cdd2787cacd89"} Nov 27 12:01:10 crc kubenswrapper[4796]: I1127 12:01:10.039784 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" podStartSLOduration=2.179707916 podStartE2EDuration="12.039762518s" podCreationTimestamp="2025-11-27 12:00:58 +0000 UTC" firstStartedPulling="2025-11-27 12:00:59.580192285 +0000 UTC m=+2177.098511203" lastFinishedPulling="2025-11-27 12:01:09.440246887 +0000 UTC m=+2186.958565805" observedRunningTime="2025-11-27 12:01:10.03470531 +0000 UTC m=+2187.553024268" watchObservedRunningTime="2025-11-27 12:01:10.039762518 +0000 UTC m=+2187.558081456" Nov 27 12:01:13 crc kubenswrapper[4796]: I1127 12:01:13.102552 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 27 12:01:13 crc kubenswrapper[4796]: I1127 12:01:13.883421 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.129739 4796 generic.go:334] "Generic (PLEG): container finished" podID="81f86614-5247-4cb9-ad8c-688c70f862a6" containerID="50725dc923c3e968338af0942bc5a8a3a49ffd705db4d1d6c58cdd2787cacd89" exitCode=0 Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.129828 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" event={"ID":"81f86614-5247-4cb9-ad8c-688c70f862a6","Type":"ContainerDied","Data":"50725dc923c3e968338af0942bc5a8a3a49ffd705db4d1d6c58cdd2787cacd89"} Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.919664 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8qns2"] Nov 27 12:01:21 crc kubenswrapper[4796]: E1127 12:01:21.920951 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48aeb330-d74c-4814-9c4d-41e40436e8a3" containerName="keystone-cron" Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.920986 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="48aeb330-d74c-4814-9c4d-41e40436e8a3" containerName="keystone-cron" Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.921514 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="48aeb330-d74c-4814-9c4d-41e40436e8a3" containerName="keystone-cron" Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.923457 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:21 crc kubenswrapper[4796]: I1127 12:01:21.942865 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8qns2"] Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.030437 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-utilities\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.030553 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ftz7\" (UniqueName: \"kubernetes.io/projected/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-kube-api-access-6ftz7\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.030617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-catalog-content\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.132747 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-utilities\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.132839 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ftz7\" (UniqueName: \"kubernetes.io/projected/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-kube-api-access-6ftz7\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.132887 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-catalog-content\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.133414 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-utilities\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.134317 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-catalog-content\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.160136 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ftz7\" (UniqueName: \"kubernetes.io/projected/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-kube-api-access-6ftz7\") pod \"community-operators-8qns2\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.257094 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.697385 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.747289 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwkm2\" (UniqueName: \"kubernetes.io/projected/81f86614-5247-4cb9-ad8c-688c70f862a6-kube-api-access-xwkm2\") pod \"81f86614-5247-4cb9-ad8c-688c70f862a6\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.747362 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-inventory\") pod \"81f86614-5247-4cb9-ad8c-688c70f862a6\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.747402 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-repo-setup-combined-ca-bundle\") pod \"81f86614-5247-4cb9-ad8c-688c70f862a6\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.747430 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-ssh-key\") pod \"81f86614-5247-4cb9-ad8c-688c70f862a6\" (UID: \"81f86614-5247-4cb9-ad8c-688c70f862a6\") " Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.754063 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f86614-5247-4cb9-ad8c-688c70f862a6-kube-api-access-xwkm2" (OuterVolumeSpecName: "kube-api-access-xwkm2") pod "81f86614-5247-4cb9-ad8c-688c70f862a6" (UID: "81f86614-5247-4cb9-ad8c-688c70f862a6"). InnerVolumeSpecName "kube-api-access-xwkm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.756496 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "81f86614-5247-4cb9-ad8c-688c70f862a6" (UID: "81f86614-5247-4cb9-ad8c-688c70f862a6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.775882 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-inventory" (OuterVolumeSpecName: "inventory") pod "81f86614-5247-4cb9-ad8c-688c70f862a6" (UID: "81f86614-5247-4cb9-ad8c-688c70f862a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.776465 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "81f86614-5247-4cb9-ad8c-688c70f862a6" (UID: "81f86614-5247-4cb9-ad8c-688c70f862a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.839514 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8qns2"] Nov 27 12:01:22 crc kubenswrapper[4796]: W1127 12:01:22.840924 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b4ca336_fe53_4d25_bacc_9dd9a400d04f.slice/crio-49a367a9371aaadbb446d74c33bafabf2b7e70389c42ffed8315d540192687a9 WatchSource:0}: Error finding container 49a367a9371aaadbb446d74c33bafabf2b7e70389c42ffed8315d540192687a9: Status 404 returned error can't find the container with id 49a367a9371aaadbb446d74c33bafabf2b7e70389c42ffed8315d540192687a9 Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.849126 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwkm2\" (UniqueName: \"kubernetes.io/projected/81f86614-5247-4cb9-ad8c-688c70f862a6-kube-api-access-xwkm2\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.849151 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.849182 4796 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:22 crc kubenswrapper[4796]: I1127 12:01:22.849192 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81f86614-5247-4cb9-ad8c-688c70f862a6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.150577 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerStarted","Data":"49a367a9371aaadbb446d74c33bafabf2b7e70389c42ffed8315d540192687a9"} Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.153176 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" event={"ID":"81f86614-5247-4cb9-ad8c-688c70f862a6","Type":"ContainerDied","Data":"9e82d6f08f61a85b86c6bbb08dbb338ba48e609e645c8b630957807c4cc97815"} Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.153209 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e82d6f08f61a85b86c6bbb08dbb338ba48e609e645c8b630957807c4cc97815" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.153338 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.236435 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6"] Nov 27 12:01:23 crc kubenswrapper[4796]: E1127 12:01:23.238897 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f86614-5247-4cb9-ad8c-688c70f862a6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.238932 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f86614-5247-4cb9-ad8c-688c70f862a6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.239221 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f86614-5247-4cb9-ad8c-688c70f862a6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.240042 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.242386 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.242932 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.243399 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.244060 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.247048 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6"] Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.359761 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.360098 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.360180 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmbgv\" (UniqueName: \"kubernetes.io/projected/9059a6c5-d130-4751-849d-dc05d1c5c037-kube-api-access-dmbgv\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.462208 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.462288 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.462340 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmbgv\" (UniqueName: \"kubernetes.io/projected/9059a6c5-d130-4751-849d-dc05d1c5c037-kube-api-access-dmbgv\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.469622 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.470076 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.482734 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmbgv\" (UniqueName: \"kubernetes.io/projected/9059a6c5-d130-4751-849d-dc05d1c5c037-kube-api-access-dmbgv\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s45b6\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:23 crc kubenswrapper[4796]: I1127 12:01:23.575703 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:24 crc kubenswrapper[4796]: I1127 12:01:24.167094 4796 generic.go:334] "Generic (PLEG): container finished" podID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerID="4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859" exitCode=0 Nov 27 12:01:24 crc kubenswrapper[4796]: I1127 12:01:24.167377 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerDied","Data":"4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859"} Nov 27 12:01:24 crc kubenswrapper[4796]: W1127 12:01:24.174904 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9059a6c5_d130_4751_849d_dc05d1c5c037.slice/crio-ab15a180395660f48b566b16536408415170623cb7dc946a19d0785988a014be WatchSource:0}: Error finding container ab15a180395660f48b566b16536408415170623cb7dc946a19d0785988a014be: Status 404 returned error can't find the container with id ab15a180395660f48b566b16536408415170623cb7dc946a19d0785988a014be Nov 27 12:01:24 crc kubenswrapper[4796]: I1127 12:01:24.180165 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6"] Nov 27 12:01:25 crc kubenswrapper[4796]: I1127 12:01:25.180312 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerStarted","Data":"722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b"} Nov 27 12:01:25 crc kubenswrapper[4796]: I1127 12:01:25.183938 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" event={"ID":"9059a6c5-d130-4751-849d-dc05d1c5c037","Type":"ContainerStarted","Data":"159a207eabe0f7763d761a1d9f711690131d50eacda800c32edd4a1cce9f757d"} Nov 27 12:01:25 crc kubenswrapper[4796]: I1127 12:01:25.183975 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" event={"ID":"9059a6c5-d130-4751-849d-dc05d1c5c037","Type":"ContainerStarted","Data":"ab15a180395660f48b566b16536408415170623cb7dc946a19d0785988a014be"} Nov 27 12:01:25 crc kubenswrapper[4796]: I1127 12:01:25.244658 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" podStartSLOduration=1.74539745 podStartE2EDuration="2.244638598s" podCreationTimestamp="2025-11-27 12:01:23 +0000 UTC" firstStartedPulling="2025-11-27 12:01:24.178318363 +0000 UTC m=+2201.696637321" lastFinishedPulling="2025-11-27 12:01:24.677559551 +0000 UTC m=+2202.195878469" observedRunningTime="2025-11-27 12:01:25.218725952 +0000 UTC m=+2202.737044920" watchObservedRunningTime="2025-11-27 12:01:25.244638598 +0000 UTC m=+2202.762957526" Nov 27 12:01:26 crc kubenswrapper[4796]: I1127 12:01:26.203632 4796 generic.go:334] "Generic (PLEG): container finished" podID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerID="722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b" exitCode=0 Nov 27 12:01:26 crc kubenswrapper[4796]: I1127 12:01:26.203759 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerDied","Data":"722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b"} Nov 27 12:01:27 crc kubenswrapper[4796]: I1127 12:01:27.217900 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerStarted","Data":"ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93"} Nov 27 12:01:27 crc kubenswrapper[4796]: I1127 12:01:27.257112 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8qns2" podStartSLOduration=3.705124002 podStartE2EDuration="6.257077563s" podCreationTimestamp="2025-11-27 12:01:21 +0000 UTC" firstStartedPulling="2025-11-27 12:01:24.170445818 +0000 UTC m=+2201.688764776" lastFinishedPulling="2025-11-27 12:01:26.722399399 +0000 UTC m=+2204.240718337" observedRunningTime="2025-11-27 12:01:27.232572785 +0000 UTC m=+2204.750891713" watchObservedRunningTime="2025-11-27 12:01:27.257077563 +0000 UTC m=+2204.775396551" Nov 27 12:01:28 crc kubenswrapper[4796]: I1127 12:01:28.230189 4796 generic.go:334] "Generic (PLEG): container finished" podID="9059a6c5-d130-4751-849d-dc05d1c5c037" containerID="159a207eabe0f7763d761a1d9f711690131d50eacda800c32edd4a1cce9f757d" exitCode=0 Nov 27 12:01:28 crc kubenswrapper[4796]: I1127 12:01:28.230253 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" event={"ID":"9059a6c5-d130-4751-849d-dc05d1c5c037","Type":"ContainerDied","Data":"159a207eabe0f7763d761a1d9f711690131d50eacda800c32edd4a1cce9f757d"} Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.702632 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.795307 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-inventory\") pod \"9059a6c5-d130-4751-849d-dc05d1c5c037\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.795763 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-ssh-key\") pod \"9059a6c5-d130-4751-849d-dc05d1c5c037\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.795876 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmbgv\" (UniqueName: \"kubernetes.io/projected/9059a6c5-d130-4751-849d-dc05d1c5c037-kube-api-access-dmbgv\") pod \"9059a6c5-d130-4751-849d-dc05d1c5c037\" (UID: \"9059a6c5-d130-4751-849d-dc05d1c5c037\") " Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.800641 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9059a6c5-d130-4751-849d-dc05d1c5c037-kube-api-access-dmbgv" (OuterVolumeSpecName: "kube-api-access-dmbgv") pod "9059a6c5-d130-4751-849d-dc05d1c5c037" (UID: "9059a6c5-d130-4751-849d-dc05d1c5c037"). InnerVolumeSpecName "kube-api-access-dmbgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.822628 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-inventory" (OuterVolumeSpecName: "inventory") pod "9059a6c5-d130-4751-849d-dc05d1c5c037" (UID: "9059a6c5-d130-4751-849d-dc05d1c5c037"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.847648 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9059a6c5-d130-4751-849d-dc05d1c5c037" (UID: "9059a6c5-d130-4751-849d-dc05d1c5c037"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.897678 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.897708 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9059a6c5-d130-4751-849d-dc05d1c5c037-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:29 crc kubenswrapper[4796]: I1127 12:01:29.897719 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmbgv\" (UniqueName: \"kubernetes.io/projected/9059a6c5-d130-4751-849d-dc05d1c5c037-kube-api-access-dmbgv\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.261969 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" event={"ID":"9059a6c5-d130-4751-849d-dc05d1c5c037","Type":"ContainerDied","Data":"ab15a180395660f48b566b16536408415170623cb7dc946a19d0785988a014be"} Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.262011 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab15a180395660f48b566b16536408415170623cb7dc946a19d0785988a014be" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.262017 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s45b6" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.347223 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx"] Nov 27 12:01:30 crc kubenswrapper[4796]: E1127 12:01:30.347613 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9059a6c5-d130-4751-849d-dc05d1c5c037" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.347627 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9059a6c5-d130-4751-849d-dc05d1c5c037" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.349482 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9059a6c5-d130-4751-849d-dc05d1c5c037" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.350129 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.354863 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.355091 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.355119 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.355201 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.367652 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx"] Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.524482 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc87g\" (UniqueName: \"kubernetes.io/projected/05691e2e-3935-4eef-9d8b-f591516ebb3c-kube-api-access-cc87g\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.524699 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.524968 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.525107 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.626816 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.626993 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc87g\" (UniqueName: \"kubernetes.io/projected/05691e2e-3935-4eef-9d8b-f591516ebb3c-kube-api-access-cc87g\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.627060 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.627179 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.630442 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.630681 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.636028 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.645139 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc87g\" (UniqueName: \"kubernetes.io/projected/05691e2e-3935-4eef-9d8b-f591516ebb3c-kube-api-access-cc87g\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:30 crc kubenswrapper[4796]: I1127 12:01:30.749931 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:01:31 crc kubenswrapper[4796]: I1127 12:01:31.257345 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx"] Nov 27 12:01:31 crc kubenswrapper[4796]: W1127 12:01:31.265152 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05691e2e_3935_4eef_9d8b_f591516ebb3c.slice/crio-59f38590e0513a69b24d222b67e58188de9f9cc6ae3f5e6ae39ecbffe2c00a24 WatchSource:0}: Error finding container 59f38590e0513a69b24d222b67e58188de9f9cc6ae3f5e6ae39ecbffe2c00a24: Status 404 returned error can't find the container with id 59f38590e0513a69b24d222b67e58188de9f9cc6ae3f5e6ae39ecbffe2c00a24 Nov 27 12:01:31 crc kubenswrapper[4796]: I1127 12:01:31.879761 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:01:31 crc kubenswrapper[4796]: I1127 12:01:31.879816 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.258601 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.258910 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.284479 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" event={"ID":"05691e2e-3935-4eef-9d8b-f591516ebb3c","Type":"ContainerStarted","Data":"1dd1edb63f06506d184bcad358c86884b30012ff2adb38f9b042d1345b4c5fc4"} Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.284522 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" event={"ID":"05691e2e-3935-4eef-9d8b-f591516ebb3c","Type":"ContainerStarted","Data":"59f38590e0513a69b24d222b67e58188de9f9cc6ae3f5e6ae39ecbffe2c00a24"} Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.307333 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" podStartSLOduration=1.873452814 podStartE2EDuration="2.30731335s" podCreationTimestamp="2025-11-27 12:01:30 +0000 UTC" firstStartedPulling="2025-11-27 12:01:31.268768522 +0000 UTC m=+2208.787087440" lastFinishedPulling="2025-11-27 12:01:31.702629058 +0000 UTC m=+2209.220947976" observedRunningTime="2025-11-27 12:01:32.300175406 +0000 UTC m=+2209.818494324" watchObservedRunningTime="2025-11-27 12:01:32.30731335 +0000 UTC m=+2209.825632278" Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.320752 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.369292 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:32 crc kubenswrapper[4796]: I1127 12:01:32.557294 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8qns2"] Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.307785 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8qns2" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="registry-server" containerID="cri-o://ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93" gracePeriod=2 Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.804304 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.814003 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ftz7\" (UniqueName: \"kubernetes.io/projected/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-kube-api-access-6ftz7\") pod \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.822496 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-kube-api-access-6ftz7" (OuterVolumeSpecName: "kube-api-access-6ftz7") pod "5b4ca336-fe53-4d25-bacc-9dd9a400d04f" (UID: "5b4ca336-fe53-4d25-bacc-9dd9a400d04f"). InnerVolumeSpecName "kube-api-access-6ftz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.915205 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-catalog-content\") pod \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.915323 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-utilities\") pod \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\" (UID: \"5b4ca336-fe53-4d25-bacc-9dd9a400d04f\") " Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.915808 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ftz7\" (UniqueName: \"kubernetes.io/projected/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-kube-api-access-6ftz7\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.916229 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-utilities" (OuterVolumeSpecName: "utilities") pod "5b4ca336-fe53-4d25-bacc-9dd9a400d04f" (UID: "5b4ca336-fe53-4d25-bacc-9dd9a400d04f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:01:34 crc kubenswrapper[4796]: I1127 12:01:34.964858 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b4ca336-fe53-4d25-bacc-9dd9a400d04f" (UID: "5b4ca336-fe53-4d25-bacc-9dd9a400d04f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.017574 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.017839 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4ca336-fe53-4d25-bacc-9dd9a400d04f-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.319798 4796 generic.go:334] "Generic (PLEG): container finished" podID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerID="ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93" exitCode=0 Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.319849 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerDied","Data":"ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93"} Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.319887 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qns2" event={"ID":"5b4ca336-fe53-4d25-bacc-9dd9a400d04f","Type":"ContainerDied","Data":"49a367a9371aaadbb446d74c33bafabf2b7e70389c42ffed8315d540192687a9"} Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.319911 4796 scope.go:117] "RemoveContainer" containerID="ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.319933 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qns2" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.341314 4796 scope.go:117] "RemoveContainer" containerID="722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.373397 4796 scope.go:117] "RemoveContainer" containerID="4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.381165 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8qns2"] Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.395422 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8qns2"] Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.439089 4796 scope.go:117] "RemoveContainer" containerID="ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93" Nov 27 12:01:35 crc kubenswrapper[4796]: E1127 12:01:35.439605 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93\": container with ID starting with ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93 not found: ID does not exist" containerID="ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.439725 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93"} err="failed to get container status \"ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93\": rpc error: code = NotFound desc = could not find container \"ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93\": container with ID starting with ac95a4d35efd23373cb132cc538f09693f624ac2bb73cd64dcd12b8eaebe3e93 not found: ID does not exist" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.439805 4796 scope.go:117] "RemoveContainer" containerID="722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b" Nov 27 12:01:35 crc kubenswrapper[4796]: E1127 12:01:35.440162 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b\": container with ID starting with 722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b not found: ID does not exist" containerID="722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.440205 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b"} err="failed to get container status \"722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b\": rpc error: code = NotFound desc = could not find container \"722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b\": container with ID starting with 722e3db4d389ab7a15b17662a23ed097fad12876aeaa9d8f95cc6965a52db46b not found: ID does not exist" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.440238 4796 scope.go:117] "RemoveContainer" containerID="4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859" Nov 27 12:01:35 crc kubenswrapper[4796]: E1127 12:01:35.440497 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859\": container with ID starting with 4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859 not found: ID does not exist" containerID="4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.440523 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859"} err="failed to get container status \"4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859\": rpc error: code = NotFound desc = could not find container \"4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859\": container with ID starting with 4e77c3bf017e746fd64794ddd1418ea46e28d4cb2f94758c137e8d7a69fd1859 not found: ID does not exist" Nov 27 12:01:35 crc kubenswrapper[4796]: I1127 12:01:35.589527 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" path="/var/lib/kubelet/pods/5b4ca336-fe53-4d25-bacc-9dd9a400d04f/volumes" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.545571 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-crldz"] Nov 27 12:01:47 crc kubenswrapper[4796]: E1127 12:01:47.546511 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="registry-server" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.546524 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="registry-server" Nov 27 12:01:47 crc kubenswrapper[4796]: E1127 12:01:47.546545 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="extract-content" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.546550 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="extract-content" Nov 27 12:01:47 crc kubenswrapper[4796]: E1127 12:01:47.546576 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="extract-utilities" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.546583 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="extract-utilities" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.546804 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4ca336-fe53-4d25-bacc-9dd9a400d04f" containerName="registry-server" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.548029 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.559163 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crldz"] Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.671353 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-utilities\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.671521 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx294\" (UniqueName: \"kubernetes.io/projected/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-kube-api-access-hx294\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.671557 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-catalog-content\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.773055 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx294\" (UniqueName: \"kubernetes.io/projected/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-kube-api-access-hx294\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.773128 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-catalog-content\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.773258 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-utilities\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.773879 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-utilities\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.773933 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-catalog-content\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.795679 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx294\" (UniqueName: \"kubernetes.io/projected/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-kube-api-access-hx294\") pod \"certified-operators-crldz\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:47 crc kubenswrapper[4796]: I1127 12:01:47.877740 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:48 crc kubenswrapper[4796]: I1127 12:01:48.367490 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crldz"] Nov 27 12:01:48 crc kubenswrapper[4796]: W1127 12:01:48.369183 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33df9cb5_3a7a_4882_93be_cf3ce37cb7b1.slice/crio-1e1a3e884b4b812395a9f5a0e2f57677e0ef53c3c82932bd995a82d559dbee8a WatchSource:0}: Error finding container 1e1a3e884b4b812395a9f5a0e2f57677e0ef53c3c82932bd995a82d559dbee8a: Status 404 returned error can't find the container with id 1e1a3e884b4b812395a9f5a0e2f57677e0ef53c3c82932bd995a82d559dbee8a Nov 27 12:01:48 crc kubenswrapper[4796]: I1127 12:01:48.468625 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crldz" event={"ID":"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1","Type":"ContainerStarted","Data":"1e1a3e884b4b812395a9f5a0e2f57677e0ef53c3c82932bd995a82d559dbee8a"} Nov 27 12:01:49 crc kubenswrapper[4796]: I1127 12:01:49.482884 4796 generic.go:334] "Generic (PLEG): container finished" podID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerID="89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416" exitCode=0 Nov 27 12:01:49 crc kubenswrapper[4796]: I1127 12:01:49.483010 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crldz" event={"ID":"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1","Type":"ContainerDied","Data":"89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416"} Nov 27 12:01:49 crc kubenswrapper[4796]: I1127 12:01:49.740366 4796 scope.go:117] "RemoveContainer" containerID="5d683f5dd13d49637f10b186e2b1c13b5c81e507a7e715a4253a8484afc9e8c6" Nov 27 12:01:51 crc kubenswrapper[4796]: I1127 12:01:51.506214 4796 generic.go:334] "Generic (PLEG): container finished" podID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerID="c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45" exitCode=0 Nov 27 12:01:51 crc kubenswrapper[4796]: I1127 12:01:51.506299 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crldz" event={"ID":"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1","Type":"ContainerDied","Data":"c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45"} Nov 27 12:01:52 crc kubenswrapper[4796]: I1127 12:01:52.535068 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crldz" event={"ID":"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1","Type":"ContainerStarted","Data":"33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3"} Nov 27 12:01:52 crc kubenswrapper[4796]: I1127 12:01:52.561817 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-crldz" podStartSLOduration=2.8055063799999997 podStartE2EDuration="5.56179817s" podCreationTimestamp="2025-11-27 12:01:47 +0000 UTC" firstStartedPulling="2025-11-27 12:01:49.488136419 +0000 UTC m=+2227.006455337" lastFinishedPulling="2025-11-27 12:01:52.244428209 +0000 UTC m=+2229.762747127" observedRunningTime="2025-11-27 12:01:52.552984399 +0000 UTC m=+2230.071303317" watchObservedRunningTime="2025-11-27 12:01:52.56179817 +0000 UTC m=+2230.080117088" Nov 27 12:01:57 crc kubenswrapper[4796]: I1127 12:01:57.878856 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:57 crc kubenswrapper[4796]: I1127 12:01:57.880589 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:57 crc kubenswrapper[4796]: I1127 12:01:57.924844 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:58 crc kubenswrapper[4796]: I1127 12:01:58.646375 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:01:58 crc kubenswrapper[4796]: I1127 12:01:58.688636 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-crldz"] Nov 27 12:02:00 crc kubenswrapper[4796]: I1127 12:02:00.611773 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-crldz" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="registry-server" containerID="cri-o://33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3" gracePeriod=2 Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.052435 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.136787 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-catalog-content\") pod \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.136929 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-utilities\") pod \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.137036 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx294\" (UniqueName: \"kubernetes.io/projected/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-kube-api-access-hx294\") pod \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\" (UID: \"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1\") " Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.138988 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-utilities" (OuterVolumeSpecName: "utilities") pod "33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" (UID: "33df9cb5-3a7a-4882-93be-cf3ce37cb7b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.146770 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-kube-api-access-hx294" (OuterVolumeSpecName: "kube-api-access-hx294") pod "33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" (UID: "33df9cb5-3a7a-4882-93be-cf3ce37cb7b1"). InnerVolumeSpecName "kube-api-access-hx294". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.238957 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx294\" (UniqueName: \"kubernetes.io/projected/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-kube-api-access-hx294\") on node \"crc\" DevicePath \"\"" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.238991 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.370307 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" (UID: "33df9cb5-3a7a-4882-93be-cf3ce37cb7b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.443378 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.623198 4796 generic.go:334] "Generic (PLEG): container finished" podID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerID="33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3" exitCode=0 Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.623291 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crldz" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.623279 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crldz" event={"ID":"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1","Type":"ContainerDied","Data":"33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3"} Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.623369 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crldz" event={"ID":"33df9cb5-3a7a-4882-93be-cf3ce37cb7b1","Type":"ContainerDied","Data":"1e1a3e884b4b812395a9f5a0e2f57677e0ef53c3c82932bd995a82d559dbee8a"} Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.623397 4796 scope.go:117] "RemoveContainer" containerID="33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.646043 4796 scope.go:117] "RemoveContainer" containerID="c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.668330 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-crldz"] Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.679390 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-crldz"] Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.679596 4796 scope.go:117] "RemoveContainer" containerID="89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.722745 4796 scope.go:117] "RemoveContainer" containerID="33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3" Nov 27 12:02:01 crc kubenswrapper[4796]: E1127 12:02:01.723112 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3\": container with ID starting with 33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3 not found: ID does not exist" containerID="33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.723154 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3"} err="failed to get container status \"33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3\": rpc error: code = NotFound desc = could not find container \"33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3\": container with ID starting with 33e8336dc710a089834ee8f065fc5110649961089212bb8b0859603b62243bd3 not found: ID does not exist" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.723175 4796 scope.go:117] "RemoveContainer" containerID="c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45" Nov 27 12:02:01 crc kubenswrapper[4796]: E1127 12:02:01.723576 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45\": container with ID starting with c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45 not found: ID does not exist" containerID="c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.723611 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45"} err="failed to get container status \"c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45\": rpc error: code = NotFound desc = could not find container \"c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45\": container with ID starting with c20f66fdd14f6b7597be3030f1d890089d9c24bfdbf9bb9d72c5b8bf114dab45 not found: ID does not exist" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.723637 4796 scope.go:117] "RemoveContainer" containerID="89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416" Nov 27 12:02:01 crc kubenswrapper[4796]: E1127 12:02:01.723960 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416\": container with ID starting with 89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416 not found: ID does not exist" containerID="89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.723994 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416"} err="failed to get container status \"89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416\": rpc error: code = NotFound desc = could not find container \"89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416\": container with ID starting with 89f86d3b565207f60bfa727e5df69bda829c7e4024864fda9809d64b97773416 not found: ID does not exist" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.879382 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.879467 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.879519 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.880244 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d02aea19dfdaf4e45b9254b7fcf0891a26f4b7262a59d6766b8fb31a94cc4ee"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:02:01 crc kubenswrapper[4796]: I1127 12:02:01.880319 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://8d02aea19dfdaf4e45b9254b7fcf0891a26f4b7262a59d6766b8fb31a94cc4ee" gracePeriod=600 Nov 27 12:02:02 crc kubenswrapper[4796]: I1127 12:02:02.639293 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="8d02aea19dfdaf4e45b9254b7fcf0891a26f4b7262a59d6766b8fb31a94cc4ee" exitCode=0 Nov 27 12:02:02 crc kubenswrapper[4796]: I1127 12:02:02.639406 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"8d02aea19dfdaf4e45b9254b7fcf0891a26f4b7262a59d6766b8fb31a94cc4ee"} Nov 27 12:02:02 crc kubenswrapper[4796]: I1127 12:02:02.639893 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f"} Nov 27 12:02:02 crc kubenswrapper[4796]: I1127 12:02:02.639929 4796 scope.go:117] "RemoveContainer" containerID="4c5cde150567087f40def3e21b4d1cdce2de67f268af21c356b295783dcc66e8" Nov 27 12:02:03 crc kubenswrapper[4796]: I1127 12:02:03.579380 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" path="/var/lib/kubelet/pods/33df9cb5-3a7a-4882-93be-cf3ce37cb7b1/volumes" Nov 27 12:02:49 crc kubenswrapper[4796]: I1127 12:02:49.816225 4796 scope.go:117] "RemoveContainer" containerID="7f6e3ea3dafac0198575955c4a149c4463bc8493e9f3101066b2fe281241530e" Nov 27 12:02:49 crc kubenswrapper[4796]: I1127 12:02:49.870306 4796 scope.go:117] "RemoveContainer" containerID="64784d397077ad3eb24b4d2ade0b7470d38a9a88edf30e8be30366a9c0a4e6d4" Nov 27 12:02:49 crc kubenswrapper[4796]: I1127 12:02:49.892763 4796 scope.go:117] "RemoveContainer" containerID="f1c6b477e6fdb687fadb6f5f399f44798a3996c714ef04a6abcd05ec70a55c89" Nov 27 12:02:49 crc kubenswrapper[4796]: I1127 12:02:49.935775 4796 scope.go:117] "RemoveContainer" containerID="f89b18a271e227ede16b294522eb260aa74c47f021952b0da1f556debaf2c649" Nov 27 12:04:31 crc kubenswrapper[4796]: I1127 12:04:31.879214 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:04:31 crc kubenswrapper[4796]: I1127 12:04:31.879977 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:04:54 crc kubenswrapper[4796]: I1127 12:04:54.404561 4796 generic.go:334] "Generic (PLEG): container finished" podID="05691e2e-3935-4eef-9d8b-f591516ebb3c" containerID="1dd1edb63f06506d184bcad358c86884b30012ff2adb38f9b042d1345b4c5fc4" exitCode=0 Nov 27 12:04:54 crc kubenswrapper[4796]: I1127 12:04:54.404645 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" event={"ID":"05691e2e-3935-4eef-9d8b-f591516ebb3c","Type":"ContainerDied","Data":"1dd1edb63f06506d184bcad358c86884b30012ff2adb38f9b042d1345b4c5fc4"} Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.853780 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.936580 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-inventory\") pod \"05691e2e-3935-4eef-9d8b-f591516ebb3c\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.936700 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc87g\" (UniqueName: \"kubernetes.io/projected/05691e2e-3935-4eef-9d8b-f591516ebb3c-kube-api-access-cc87g\") pod \"05691e2e-3935-4eef-9d8b-f591516ebb3c\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.936889 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-ssh-key\") pod \"05691e2e-3935-4eef-9d8b-f591516ebb3c\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.937009 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-bootstrap-combined-ca-bundle\") pod \"05691e2e-3935-4eef-9d8b-f591516ebb3c\" (UID: \"05691e2e-3935-4eef-9d8b-f591516ebb3c\") " Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.949789 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "05691e2e-3935-4eef-9d8b-f591516ebb3c" (UID: "05691e2e-3935-4eef-9d8b-f591516ebb3c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.949859 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05691e2e-3935-4eef-9d8b-f591516ebb3c-kube-api-access-cc87g" (OuterVolumeSpecName: "kube-api-access-cc87g") pod "05691e2e-3935-4eef-9d8b-f591516ebb3c" (UID: "05691e2e-3935-4eef-9d8b-f591516ebb3c"). InnerVolumeSpecName "kube-api-access-cc87g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.981193 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "05691e2e-3935-4eef-9d8b-f591516ebb3c" (UID: "05691e2e-3935-4eef-9d8b-f591516ebb3c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:04:55 crc kubenswrapper[4796]: I1127 12:04:55.988500 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-inventory" (OuterVolumeSpecName: "inventory") pod "05691e2e-3935-4eef-9d8b-f591516ebb3c" (UID: "05691e2e-3935-4eef-9d8b-f591516ebb3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.040904 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.041462 4796 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.041479 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05691e2e-3935-4eef-9d8b-f591516ebb3c-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.041493 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc87g\" (UniqueName: \"kubernetes.io/projected/05691e2e-3935-4eef-9d8b-f591516ebb3c-kube-api-access-cc87g\") on node \"crc\" DevicePath \"\"" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.423086 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" event={"ID":"05691e2e-3935-4eef-9d8b-f591516ebb3c","Type":"ContainerDied","Data":"59f38590e0513a69b24d222b67e58188de9f9cc6ae3f5e6ae39ecbffe2c00a24"} Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.423127 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f38590e0513a69b24d222b67e58188de9f9cc6ae3f5e6ae39ecbffe2c00a24" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.423155 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.505536 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm"] Nov 27 12:04:56 crc kubenswrapper[4796]: E1127 12:04:56.506030 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="extract-utilities" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.506053 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="extract-utilities" Nov 27 12:04:56 crc kubenswrapper[4796]: E1127 12:04:56.506067 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="extract-content" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.506075 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="extract-content" Nov 27 12:04:56 crc kubenswrapper[4796]: E1127 12:04:56.506094 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05691e2e-3935-4eef-9d8b-f591516ebb3c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.506104 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="05691e2e-3935-4eef-9d8b-f591516ebb3c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 12:04:56 crc kubenswrapper[4796]: E1127 12:04:56.506128 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="registry-server" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.506136 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="registry-server" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.506421 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="33df9cb5-3a7a-4882-93be-cf3ce37cb7b1" containerName="registry-server" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.506447 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="05691e2e-3935-4eef-9d8b-f591516ebb3c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.507184 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.513805 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.513859 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.514063 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.522251 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.546741 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm"] Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.553741 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.553792 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.553816 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd9h9\" (UniqueName: \"kubernetes.io/projected/72e23875-5131-453e-a9d5-dcdbea0b4cf9-kube-api-access-xd9h9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.655255 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.656095 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.656125 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd9h9\" (UniqueName: \"kubernetes.io/projected/72e23875-5131-453e-a9d5-dcdbea0b4cf9-kube-api-access-xd9h9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.660447 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.661134 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.675154 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd9h9\" (UniqueName: \"kubernetes.io/projected/72e23875-5131-453e-a9d5-dcdbea0b4cf9-kube-api-access-xd9h9\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:56 crc kubenswrapper[4796]: I1127 12:04:56.843906 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:04:57 crc kubenswrapper[4796]: I1127 12:04:57.437216 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm"] Nov 27 12:04:58 crc kubenswrapper[4796]: I1127 12:04:58.453068 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" event={"ID":"72e23875-5131-453e-a9d5-dcdbea0b4cf9","Type":"ContainerStarted","Data":"18d18103d5713bd945d7c7a5144585a76b733796d844079c53dac5c1c1ae945c"} Nov 27 12:04:58 crc kubenswrapper[4796]: I1127 12:04:58.453370 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" event={"ID":"72e23875-5131-453e-a9d5-dcdbea0b4cf9","Type":"ContainerStarted","Data":"e2d8814c8e786776bf156b2e8fd5281718992fdc6d161382425cef565c054bc6"} Nov 27 12:04:58 crc kubenswrapper[4796]: I1127 12:04:58.479974 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" podStartSLOduration=1.7194206680000002 podStartE2EDuration="2.479956003s" podCreationTimestamp="2025-11-27 12:04:56 +0000 UTC" firstStartedPulling="2025-11-27 12:04:57.429590285 +0000 UTC m=+2414.947909213" lastFinishedPulling="2025-11-27 12:04:58.19012563 +0000 UTC m=+2415.708444548" observedRunningTime="2025-11-27 12:04:58.467840803 +0000 UTC m=+2415.986159721" watchObservedRunningTime="2025-11-27 12:04:58.479956003 +0000 UTC m=+2415.998274911" Nov 27 12:05:01 crc kubenswrapper[4796]: I1127 12:05:01.879499 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:05:01 crc kubenswrapper[4796]: I1127 12:05:01.880040 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.708750 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t99f4"] Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.711161 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.718413 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t99f4"] Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.749841 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsqb4\" (UniqueName: \"kubernetes.io/projected/98b2c708-f573-4d0d-b757-90217c0e633b-kube-api-access-tsqb4\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.750001 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-utilities\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.750382 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-catalog-content\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.852454 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-utilities\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.852589 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-catalog-content\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.852689 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsqb4\" (UniqueName: \"kubernetes.io/projected/98b2c708-f573-4d0d-b757-90217c0e633b-kube-api-access-tsqb4\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.852905 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-utilities\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.852951 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-catalog-content\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.885842 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsqb4\" (UniqueName: \"kubernetes.io/projected/98b2c708-f573-4d0d-b757-90217c0e633b-kube-api-access-tsqb4\") pod \"redhat-marketplace-t99f4\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.948532 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6mtqq"] Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.960684 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:25 crc kubenswrapper[4796]: I1127 12:05:25.979766 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mtqq"] Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.056874 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-utilities\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.056927 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-catalog-content\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.057089 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vpz2\" (UniqueName: \"kubernetes.io/projected/ef2bad3a-6372-484e-8107-2cec57c1b3a3-kube-api-access-4vpz2\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.068508 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.158454 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vpz2\" (UniqueName: \"kubernetes.io/projected/ef2bad3a-6372-484e-8107-2cec57c1b3a3-kube-api-access-4vpz2\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.158519 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-utilities\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.158568 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-catalog-content\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.159169 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-catalog-content\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.160245 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-utilities\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.188776 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vpz2\" (UniqueName: \"kubernetes.io/projected/ef2bad3a-6372-484e-8107-2cec57c1b3a3-kube-api-access-4vpz2\") pod \"redhat-operators-6mtqq\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.289543 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.464827 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t99f4"] Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.702957 4796 generic.go:334] "Generic (PLEG): container finished" podID="98b2c708-f573-4d0d-b757-90217c0e633b" containerID="5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6" exitCode=0 Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.703003 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t99f4" event={"ID":"98b2c708-f573-4d0d-b757-90217c0e633b","Type":"ContainerDied","Data":"5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6"} Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.703230 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t99f4" event={"ID":"98b2c708-f573-4d0d-b757-90217c0e633b","Type":"ContainerStarted","Data":"5cb0436949c63453aa92f8f8a9d58e923809257b79ddf8eaee69ae5bed2d7bdc"} Nov 27 12:05:26 crc kubenswrapper[4796]: W1127 12:05:26.819202 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef2bad3a_6372_484e_8107_2cec57c1b3a3.slice/crio-edf60fbd348818cc2aa330050448886f539d67886e98c636d23bc1eea059f2a6 WatchSource:0}: Error finding container edf60fbd348818cc2aa330050448886f539d67886e98c636d23bc1eea059f2a6: Status 404 returned error can't find the container with id edf60fbd348818cc2aa330050448886f539d67886e98c636d23bc1eea059f2a6 Nov 27 12:05:26 crc kubenswrapper[4796]: I1127 12:05:26.819613 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mtqq"] Nov 27 12:05:27 crc kubenswrapper[4796]: I1127 12:05:27.718799 4796 generic.go:334] "Generic (PLEG): container finished" podID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerID="8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961" exitCode=0 Nov 27 12:05:27 crc kubenswrapper[4796]: I1127 12:05:27.719056 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerDied","Data":"8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961"} Nov 27 12:05:27 crc kubenswrapper[4796]: I1127 12:05:27.719081 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerStarted","Data":"edf60fbd348818cc2aa330050448886f539d67886e98c636d23bc1eea059f2a6"} Nov 27 12:05:28 crc kubenswrapper[4796]: I1127 12:05:28.728457 4796 generic.go:334] "Generic (PLEG): container finished" podID="98b2c708-f573-4d0d-b757-90217c0e633b" containerID="4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8" exitCode=0 Nov 27 12:05:28 crc kubenswrapper[4796]: I1127 12:05:28.728549 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t99f4" event={"ID":"98b2c708-f573-4d0d-b757-90217c0e633b","Type":"ContainerDied","Data":"4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8"} Nov 27 12:05:29 crc kubenswrapper[4796]: I1127 12:05:29.746777 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerStarted","Data":"8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d"} Nov 27 12:05:30 crc kubenswrapper[4796]: I1127 12:05:30.758717 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t99f4" event={"ID":"98b2c708-f573-4d0d-b757-90217c0e633b","Type":"ContainerStarted","Data":"5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb"} Nov 27 12:05:30 crc kubenswrapper[4796]: I1127 12:05:30.790007 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t99f4" podStartSLOduration=2.437938983 podStartE2EDuration="5.789990176s" podCreationTimestamp="2025-11-27 12:05:25 +0000 UTC" firstStartedPulling="2025-11-27 12:05:26.704431174 +0000 UTC m=+2444.222757053" lastFinishedPulling="2025-11-27 12:05:30.056489328 +0000 UTC m=+2447.574808246" observedRunningTime="2025-11-27 12:05:30.785504123 +0000 UTC m=+2448.303823041" watchObservedRunningTime="2025-11-27 12:05:30.789990176 +0000 UTC m=+2448.308309094" Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.773094 4796 generic.go:334] "Generic (PLEG): container finished" podID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerID="8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d" exitCode=0 Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.773190 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerDied","Data":"8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d"} Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.878756 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.878827 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.878886 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.879771 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:05:31 crc kubenswrapper[4796]: I1127 12:05:31.879853 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" gracePeriod=600 Nov 27 12:05:32 crc kubenswrapper[4796]: E1127 12:05:32.195072 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:05:32 crc kubenswrapper[4796]: I1127 12:05:32.789469 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" exitCode=0 Nov 27 12:05:32 crc kubenswrapper[4796]: I1127 12:05:32.789520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f"} Nov 27 12:05:32 crc kubenswrapper[4796]: I1127 12:05:32.789558 4796 scope.go:117] "RemoveContainer" containerID="8d02aea19dfdaf4e45b9254b7fcf0891a26f4b7262a59d6766b8fb31a94cc4ee" Nov 27 12:05:32 crc kubenswrapper[4796]: I1127 12:05:32.790563 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:05:32 crc kubenswrapper[4796]: E1127 12:05:32.791597 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:05:33 crc kubenswrapper[4796]: I1127 12:05:33.803592 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerStarted","Data":"b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177"} Nov 27 12:05:33 crc kubenswrapper[4796]: I1127 12:05:33.824459 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6mtqq" podStartSLOduration=3.624171261 podStartE2EDuration="8.824441117s" podCreationTimestamp="2025-11-27 12:05:25 +0000 UTC" firstStartedPulling="2025-11-27 12:05:27.720887388 +0000 UTC m=+2445.239206306" lastFinishedPulling="2025-11-27 12:05:32.921157244 +0000 UTC m=+2450.439476162" observedRunningTime="2025-11-27 12:05:33.824440627 +0000 UTC m=+2451.342759545" watchObservedRunningTime="2025-11-27 12:05:33.824441117 +0000 UTC m=+2451.342760035" Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.044196 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9d19-account-create-update-f7228"] Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.056024 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cbac-account-create-update-jn7bg"] Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.069571 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.069644 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.070346 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-7shzf"] Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.081713 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-7shzf"] Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.087797 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-cbac-account-create-update-jn7bg"] Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.096001 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9d19-account-create-update-f7228"] Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.122813 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.291379 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.291431 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:36 crc kubenswrapper[4796]: I1127 12:05:36.906306 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.028621 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b679-account-create-update-pbbpm"] Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.038296 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-hxhjx"] Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.049299 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ngbvt"] Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.057132 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b679-account-create-update-pbbpm"] Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.065384 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-hxhjx"] Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.076710 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ngbvt"] Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.359036 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6mtqq" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="registry-server" probeResult="failure" output=< Nov 27 12:05:37 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:05:37 crc kubenswrapper[4796]: > Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.583840 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25cf6b67-ab55-4aa8-a2d6-40421ec20aa6" path="/var/lib/kubelet/pods/25cf6b67-ab55-4aa8-a2d6-40421ec20aa6/volumes" Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.584815 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="407847a5-70c0-43f0-b86f-3b5d6e64ff22" path="/var/lib/kubelet/pods/407847a5-70c0-43f0-b86f-3b5d6e64ff22/volumes" Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.585727 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cab3a5d-5bee-478c-9597-6742f2357e86" path="/var/lib/kubelet/pods/6cab3a5d-5bee-478c-9597-6742f2357e86/volumes" Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.586645 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db284c6-3140-4359-b58b-e3868608590a" path="/var/lib/kubelet/pods/7db284c6-3140-4359-b58b-e3868608590a/volumes" Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.588472 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7b76a35-a5ea-4a21-9858-a48f47b6bead" path="/var/lib/kubelet/pods/c7b76a35-a5ea-4a21-9858-a48f47b6bead/volumes" Nov 27 12:05:37 crc kubenswrapper[4796]: I1127 12:05:37.589051 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fff7060c-0cf3-49e4-811a-aeaf656601a0" path="/var/lib/kubelet/pods/fff7060c-0cf3-49e4-811a-aeaf656601a0/volumes" Nov 27 12:05:39 crc kubenswrapper[4796]: I1127 12:05:39.900020 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t99f4"] Nov 27 12:05:39 crc kubenswrapper[4796]: I1127 12:05:39.900714 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t99f4" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="registry-server" containerID="cri-o://5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb" gracePeriod=2 Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.363451 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.543395 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-utilities\") pod \"98b2c708-f573-4d0d-b757-90217c0e633b\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.543486 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsqb4\" (UniqueName: \"kubernetes.io/projected/98b2c708-f573-4d0d-b757-90217c0e633b-kube-api-access-tsqb4\") pod \"98b2c708-f573-4d0d-b757-90217c0e633b\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.543696 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-catalog-content\") pod \"98b2c708-f573-4d0d-b757-90217c0e633b\" (UID: \"98b2c708-f573-4d0d-b757-90217c0e633b\") " Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.544380 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-utilities" (OuterVolumeSpecName: "utilities") pod "98b2c708-f573-4d0d-b757-90217c0e633b" (UID: "98b2c708-f573-4d0d-b757-90217c0e633b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.551520 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98b2c708-f573-4d0d-b757-90217c0e633b-kube-api-access-tsqb4" (OuterVolumeSpecName: "kube-api-access-tsqb4") pod "98b2c708-f573-4d0d-b757-90217c0e633b" (UID: "98b2c708-f573-4d0d-b757-90217c0e633b"). InnerVolumeSpecName "kube-api-access-tsqb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.573359 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98b2c708-f573-4d0d-b757-90217c0e633b" (UID: "98b2c708-f573-4d0d-b757-90217c0e633b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.646757 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.646796 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b2c708-f573-4d0d-b757-90217c0e633b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.646822 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsqb4\" (UniqueName: \"kubernetes.io/projected/98b2c708-f573-4d0d-b757-90217c0e633b-kube-api-access-tsqb4\") on node \"crc\" DevicePath \"\"" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.898211 4796 generic.go:334] "Generic (PLEG): container finished" podID="98b2c708-f573-4d0d-b757-90217c0e633b" containerID="5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb" exitCode=0 Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.898304 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t99f4" event={"ID":"98b2c708-f573-4d0d-b757-90217c0e633b","Type":"ContainerDied","Data":"5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb"} Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.898356 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t99f4" event={"ID":"98b2c708-f573-4d0d-b757-90217c0e633b","Type":"ContainerDied","Data":"5cb0436949c63453aa92f8f8a9d58e923809257b79ddf8eaee69ae5bed2d7bdc"} Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.898354 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t99f4" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.898390 4796 scope.go:117] "RemoveContainer" containerID="5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.936386 4796 scope.go:117] "RemoveContainer" containerID="4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8" Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.957599 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t99f4"] Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.968255 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t99f4"] Nov 27 12:05:40 crc kubenswrapper[4796]: I1127 12:05:40.971821 4796 scope.go:117] "RemoveContainer" containerID="5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.045509 4796 scope.go:117] "RemoveContainer" containerID="5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb" Nov 27 12:05:41 crc kubenswrapper[4796]: E1127 12:05:41.046219 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb\": container with ID starting with 5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb not found: ID does not exist" containerID="5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.046329 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb"} err="failed to get container status \"5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb\": rpc error: code = NotFound desc = could not find container \"5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb\": container with ID starting with 5b865f4c00d45c1cd66be067ed4510ed05edf402fe90b2604199a7db8dfd94cb not found: ID does not exist" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.046376 4796 scope.go:117] "RemoveContainer" containerID="4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8" Nov 27 12:05:41 crc kubenswrapper[4796]: E1127 12:05:41.046952 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8\": container with ID starting with 4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8 not found: ID does not exist" containerID="4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.047002 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8"} err="failed to get container status \"4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8\": rpc error: code = NotFound desc = could not find container \"4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8\": container with ID starting with 4bc814042bdb97cf2e79986930ebf6e3a3089da9b95666184253dcf5bd8b81e8 not found: ID does not exist" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.047030 4796 scope.go:117] "RemoveContainer" containerID="5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6" Nov 27 12:05:41 crc kubenswrapper[4796]: E1127 12:05:41.047481 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6\": container with ID starting with 5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6 not found: ID does not exist" containerID="5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.047533 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6"} err="failed to get container status \"5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6\": rpc error: code = NotFound desc = could not find container \"5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6\": container with ID starting with 5aee64a26ebc2b8fd82664b636e6ed4a627fcbf03e439185efa754d2a523c2d6 not found: ID does not exist" Nov 27 12:05:41 crc kubenswrapper[4796]: I1127 12:05:41.583127 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" path="/var/lib/kubelet/pods/98b2c708-f573-4d0d-b757-90217c0e633b/volumes" Nov 27 12:05:44 crc kubenswrapper[4796]: I1127 12:05:44.569472 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:05:44 crc kubenswrapper[4796]: E1127 12:05:44.570080 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:05:46 crc kubenswrapper[4796]: I1127 12:05:46.346425 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:46 crc kubenswrapper[4796]: I1127 12:05:46.413959 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:46 crc kubenswrapper[4796]: I1127 12:05:46.587727 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6mtqq"] Nov 27 12:05:47 crc kubenswrapper[4796]: I1127 12:05:47.972118 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6mtqq" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="registry-server" containerID="cri-o://b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177" gracePeriod=2 Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.437064 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.635471 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-catalog-content\") pod \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.635856 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vpz2\" (UniqueName: \"kubernetes.io/projected/ef2bad3a-6372-484e-8107-2cec57c1b3a3-kube-api-access-4vpz2\") pod \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.636052 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-utilities\") pod \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\" (UID: \"ef2bad3a-6372-484e-8107-2cec57c1b3a3\") " Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.637209 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-utilities" (OuterVolumeSpecName: "utilities") pod "ef2bad3a-6372-484e-8107-2cec57c1b3a3" (UID: "ef2bad3a-6372-484e-8107-2cec57c1b3a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.641934 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2bad3a-6372-484e-8107-2cec57c1b3a3-kube-api-access-4vpz2" (OuterVolumeSpecName: "kube-api-access-4vpz2") pod "ef2bad3a-6372-484e-8107-2cec57c1b3a3" (UID: "ef2bad3a-6372-484e-8107-2cec57c1b3a3"). InnerVolumeSpecName "kube-api-access-4vpz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.722970 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef2bad3a-6372-484e-8107-2cec57c1b3a3" (UID: "ef2bad3a-6372-484e-8107-2cec57c1b3a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.738839 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.739531 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vpz2\" (UniqueName: \"kubernetes.io/projected/ef2bad3a-6372-484e-8107-2cec57c1b3a3-kube-api-access-4vpz2\") on node \"crc\" DevicePath \"\"" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.739553 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2bad3a-6372-484e-8107-2cec57c1b3a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.985627 4796 generic.go:334] "Generic (PLEG): container finished" podID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerID="b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177" exitCode=0 Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.985694 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerDied","Data":"b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177"} Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.985736 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mtqq" Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.985779 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mtqq" event={"ID":"ef2bad3a-6372-484e-8107-2cec57c1b3a3","Type":"ContainerDied","Data":"edf60fbd348818cc2aa330050448886f539d67886e98c636d23bc1eea059f2a6"} Nov 27 12:05:48 crc kubenswrapper[4796]: I1127 12:05:48.985805 4796 scope.go:117] "RemoveContainer" containerID="b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.028356 4796 scope.go:117] "RemoveContainer" containerID="8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.030527 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6mtqq"] Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.054379 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6mtqq"] Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.056091 4796 scope.go:117] "RemoveContainer" containerID="8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.119254 4796 scope.go:117] "RemoveContainer" containerID="b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177" Nov 27 12:05:49 crc kubenswrapper[4796]: E1127 12:05:49.119730 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177\": container with ID starting with b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177 not found: ID does not exist" containerID="b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.119768 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177"} err="failed to get container status \"b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177\": rpc error: code = NotFound desc = could not find container \"b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177\": container with ID starting with b9347f42a336985af894a7e816011c95772ce75720634ae6bb06e969e5a82177 not found: ID does not exist" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.119794 4796 scope.go:117] "RemoveContainer" containerID="8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d" Nov 27 12:05:49 crc kubenswrapper[4796]: E1127 12:05:49.120137 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d\": container with ID starting with 8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d not found: ID does not exist" containerID="8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.120169 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d"} err="failed to get container status \"8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d\": rpc error: code = NotFound desc = could not find container \"8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d\": container with ID starting with 8db7c812371063096de39e594e836e99dd850538b238c07b792d04fdd552778d not found: ID does not exist" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.120192 4796 scope.go:117] "RemoveContainer" containerID="8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961" Nov 27 12:05:49 crc kubenswrapper[4796]: E1127 12:05:49.120508 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961\": container with ID starting with 8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961 not found: ID does not exist" containerID="8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.120537 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961"} err="failed to get container status \"8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961\": rpc error: code = NotFound desc = could not find container \"8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961\": container with ID starting with 8ec6f77b309e6a7eb175196759f9ee70943c6ede1a899dfa5cbac7e4499f7961 not found: ID does not exist" Nov 27 12:05:49 crc kubenswrapper[4796]: I1127 12:05:49.582732 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" path="/var/lib/kubelet/pods/ef2bad3a-6372-484e-8107-2cec57c1b3a3/volumes" Nov 27 12:05:50 crc kubenswrapper[4796]: I1127 12:05:50.202502 4796 scope.go:117] "RemoveContainer" containerID="72504b6c88ea44cf28e7a64fb9ac76c726e04b938d99f1c9bb579376098afa80" Nov 27 12:05:50 crc kubenswrapper[4796]: I1127 12:05:50.229678 4796 scope.go:117] "RemoveContainer" containerID="3a365ff88e03082ed880b00386d400b9bc2970ac2bab6e65494c89318ce4c010" Nov 27 12:05:50 crc kubenswrapper[4796]: I1127 12:05:50.299617 4796 scope.go:117] "RemoveContainer" containerID="0c021bce9cd8479a5fb869b93b1bba82be398e2a123aef602dd5993947771b65" Nov 27 12:05:50 crc kubenswrapper[4796]: I1127 12:05:50.352249 4796 scope.go:117] "RemoveContainer" containerID="5217d4919a1f79de0a525fac7d36476a70e4cbf6ea33327718d60edb741d9991" Nov 27 12:05:50 crc kubenswrapper[4796]: I1127 12:05:50.398470 4796 scope.go:117] "RemoveContainer" containerID="1032d80608d2b61125c35d5d4f01c35936ecfc6db90403990e6b0e05b996472a" Nov 27 12:05:50 crc kubenswrapper[4796]: I1127 12:05:50.449042 4796 scope.go:117] "RemoveContainer" containerID="51f943c70a0403d361fa6f7378b63cd404a0c9a667f2c29329c5fe5612b031cc" Nov 27 12:05:58 crc kubenswrapper[4796]: I1127 12:05:58.568836 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:05:58 crc kubenswrapper[4796]: E1127 12:05:58.569688 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:06:09 crc kubenswrapper[4796]: I1127 12:06:09.043074 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-rc66k"] Nov 27 12:06:09 crc kubenswrapper[4796]: I1127 12:06:09.051719 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-rc66k"] Nov 27 12:06:09 crc kubenswrapper[4796]: I1127 12:06:09.582413 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79e9fbbf-a929-4d4a-8fb9-ab083c0aea21" path="/var/lib/kubelet/pods/79e9fbbf-a929-4d4a-8fb9-ab083c0aea21/volumes" Nov 27 12:06:10 crc kubenswrapper[4796]: I1127 12:06:10.569883 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:06:10 crc kubenswrapper[4796]: E1127 12:06:10.570225 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:06:11 crc kubenswrapper[4796]: I1127 12:06:11.034054 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xr4s2"] Nov 27 12:06:11 crc kubenswrapper[4796]: I1127 12:06:11.046004 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xr4s2"] Nov 27 12:06:11 crc kubenswrapper[4796]: I1127 12:06:11.579074 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="430e7010-722c-4c22-9097-b65ed2975115" path="/var/lib/kubelet/pods/430e7010-722c-4c22-9097-b65ed2975115/volumes" Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.040091 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-23af-account-create-update-4vg5g"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.048430 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-25q46"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.057343 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-m7gr7"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.067760 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f34e-account-create-update-w5g78"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.077450 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-622c8"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.085274 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-m7gr7"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.093095 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a1e4-account-create-update-5hjzz"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.125670 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-77aa-account-create-update-qqwnn"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.135476 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f34e-account-create-update-w5g78"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.146050 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-23af-account-create-update-4vg5g"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.156742 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-25q46"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.166106 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-77aa-account-create-update-qqwnn"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.178097 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-622c8"] Nov 27 12:06:16 crc kubenswrapper[4796]: I1127 12:06:16.187370 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a1e4-account-create-update-5hjzz"] Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.579511 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="590cb807-ad51-4176-a32c-0366c5d976c5" path="/var/lib/kubelet/pods/590cb807-ad51-4176-a32c-0366c5d976c5/volumes" Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.580915 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85e62fd9-c328-49cf-8f64-a61fb678a313" path="/var/lib/kubelet/pods/85e62fd9-c328-49cf-8f64-a61fb678a313/volumes" Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.581643 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9938c442-847a-48f5-bcd4-7454e8a7c7e1" path="/var/lib/kubelet/pods/9938c442-847a-48f5-bcd4-7454e8a7c7e1/volumes" Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.582380 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c223cd8-b92c-470b-bf11-f0c41d352887" path="/var/lib/kubelet/pods/9c223cd8-b92c-470b-bf11-f0c41d352887/volumes" Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.583612 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e84a3d5c-420f-437c-9ba9-901f550d54a7" path="/var/lib/kubelet/pods/e84a3d5c-420f-437c-9ba9-901f550d54a7/volumes" Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.584326 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ac8842-ecad-4b5d-a7e9-6daa85540618" path="/var/lib/kubelet/pods/e8ac8842-ecad-4b5d-a7e9-6daa85540618/volumes" Nov 27 12:06:17 crc kubenswrapper[4796]: I1127 12:06:17.585013 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc65938f-ce7e-462d-bd24-b8265723c981" path="/var/lib/kubelet/pods/fc65938f-ce7e-462d-bd24-b8265723c981/volumes" Nov 27 12:06:20 crc kubenswrapper[4796]: I1127 12:06:20.030260 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-qgs76"] Nov 27 12:06:20 crc kubenswrapper[4796]: I1127 12:06:20.040444 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-qgs76"] Nov 27 12:06:21 crc kubenswrapper[4796]: I1127 12:06:21.569230 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:06:21 crc kubenswrapper[4796]: E1127 12:06:21.569854 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:06:21 crc kubenswrapper[4796]: I1127 12:06:21.579116 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9833ccd1-3ec0-487a-9ba3-b536a7f25829" path="/var/lib/kubelet/pods/9833ccd1-3ec0-487a-9ba3-b536a7f25829/volumes" Nov 27 12:06:28 crc kubenswrapper[4796]: I1127 12:06:28.467002 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-drlr5" podUID="5865e45d-8c50-45ea-9fb5-f8a14385de1a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.81:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 12:06:28 crc kubenswrapper[4796]: I1127 12:06:28.479981 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-wkg8j" podUID="24bc27f0-611d-4147-9819-6ec0eb012d81" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 12:06:35 crc kubenswrapper[4796]: I1127 12:06:35.569132 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:06:35 crc kubenswrapper[4796]: E1127 12:06:35.570320 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:06:48 crc kubenswrapper[4796]: I1127 12:06:48.569491 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:06:48 crc kubenswrapper[4796]: E1127 12:06:48.570337 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:06:49 crc kubenswrapper[4796]: I1127 12:06:49.043215 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-bbvsz"] Nov 27 12:06:49 crc kubenswrapper[4796]: I1127 12:06:49.051320 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-bbvsz"] Nov 27 12:06:49 crc kubenswrapper[4796]: I1127 12:06:49.581463 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4924cd5e-791b-4312-b9bc-ad1167ec5fe4" path="/var/lib/kubelet/pods/4924cd5e-791b-4312-b9bc-ad1167ec5fe4/volumes" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.686955 4796 scope.go:117] "RemoveContainer" containerID="b1d7416c65f885ca7a9f5fae131a4a5868e7716faec7f0727c62b2869e78ccee" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.726158 4796 scope.go:117] "RemoveContainer" containerID="05f28f23c9d44321b0d441bb0983ccfa294b974fe75a951c9b4218acf1018a51" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.768658 4796 scope.go:117] "RemoveContainer" containerID="364ff08d502464b849bd6befb4c6bded68e2c1e9ecd53e9a77f5d0f7b493c170" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.814342 4796 scope.go:117] "RemoveContainer" containerID="7741b2af51aca03e3d3d35359f0ead06621672ac702eaadaa751bb52ed132e4d" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.853470 4796 scope.go:117] "RemoveContainer" containerID="4c8f5b43130fd3a0ee0a2575e5c9f40c418136a30c16b319e0eeff5e7f5043ac" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.934543 4796 scope.go:117] "RemoveContainer" containerID="3817b175de387b9b45e2967aac4ebb34c0f8e5a6a6a82648b0bacf995423d79a" Nov 27 12:06:50 crc kubenswrapper[4796]: I1127 12:06:50.978873 4796 scope.go:117] "RemoveContainer" containerID="75892ccedad2f51d2021de4b3d7ab47f62dac7c2e87e1399591fc083cba8ea67" Nov 27 12:06:51 crc kubenswrapper[4796]: I1127 12:06:51.003021 4796 scope.go:117] "RemoveContainer" containerID="b578ec53d5f500a9fb7067ffdeed3cd4222a51aa161b02a1f7f705fa980b5a12" Nov 27 12:06:51 crc kubenswrapper[4796]: I1127 12:06:51.022495 4796 scope.go:117] "RemoveContainer" containerID="7774f31c04e4de76576e1f215c57f1bca6c0ac373466f09fdd67a64c9d2f0b45" Nov 27 12:06:51 crc kubenswrapper[4796]: I1127 12:06:51.047409 4796 scope.go:117] "RemoveContainer" containerID="93aab9689a1f5e7f1df961b11a2fad20b297a011daa2152155e216b659ea760d" Nov 27 12:06:51 crc kubenswrapper[4796]: I1127 12:06:51.071194 4796 scope.go:117] "RemoveContainer" containerID="deb31ef07e4a5dfc949da1fc2b761a3a9035fcbff0cf81587d0b972e4e1681a5" Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.035330 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dwbsz"] Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.045855 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dwbsz"] Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.055441 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-svvgz"] Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.065372 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-svvgz"] Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.580052 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6716271b-432f-4395-8d1a-3915979f3c33" path="/var/lib/kubelet/pods/6716271b-432f-4395-8d1a-3915979f3c33/volumes" Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.580680 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3" path="/var/lib/kubelet/pods/f8d6ccfb-6b49-4e63-b4fe-2e061f4fc6e3/volumes" Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.850050 4796 generic.go:334] "Generic (PLEG): container finished" podID="72e23875-5131-453e-a9d5-dcdbea0b4cf9" containerID="18d18103d5713bd945d7c7a5144585a76b733796d844079c53dac5c1c1ae945c" exitCode=0 Nov 27 12:06:59 crc kubenswrapper[4796]: I1127 12:06:59.850094 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" event={"ID":"72e23875-5131-453e-a9d5-dcdbea0b4cf9","Type":"ContainerDied","Data":"18d18103d5713bd945d7c7a5144585a76b733796d844079c53dac5c1c1ae945c"} Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.339972 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.491479 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd9h9\" (UniqueName: \"kubernetes.io/projected/72e23875-5131-453e-a9d5-dcdbea0b4cf9-kube-api-access-xd9h9\") pod \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.491641 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-ssh-key\") pod \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.491683 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-inventory\") pod \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\" (UID: \"72e23875-5131-453e-a9d5-dcdbea0b4cf9\") " Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.498104 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e23875-5131-453e-a9d5-dcdbea0b4cf9-kube-api-access-xd9h9" (OuterVolumeSpecName: "kube-api-access-xd9h9") pod "72e23875-5131-453e-a9d5-dcdbea0b4cf9" (UID: "72e23875-5131-453e-a9d5-dcdbea0b4cf9"). InnerVolumeSpecName "kube-api-access-xd9h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.521343 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-inventory" (OuterVolumeSpecName: "inventory") pod "72e23875-5131-453e-a9d5-dcdbea0b4cf9" (UID: "72e23875-5131-453e-a9d5-dcdbea0b4cf9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.535279 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "72e23875-5131-453e-a9d5-dcdbea0b4cf9" (UID: "72e23875-5131-453e-a9d5-dcdbea0b4cf9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.594108 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd9h9\" (UniqueName: \"kubernetes.io/projected/72e23875-5131-453e-a9d5-dcdbea0b4cf9-kube-api-access-xd9h9\") on node \"crc\" DevicePath \"\"" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.594139 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.594150 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e23875-5131-453e-a9d5-dcdbea0b4cf9-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.882116 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" event={"ID":"72e23875-5131-453e-a9d5-dcdbea0b4cf9","Type":"ContainerDied","Data":"e2d8814c8e786776bf156b2e8fd5281718992fdc6d161382425cef565c054bc6"} Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.882172 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2d8814c8e786776bf156b2e8fd5281718992fdc6d161382425cef565c054bc6" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.882169 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960192 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d"] Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960627 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="extract-utilities" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960650 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="extract-utilities" Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960670 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="extract-content" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960677 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="extract-content" Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960691 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="registry-server" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960697 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="registry-server" Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960713 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="extract-utilities" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960720 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="extract-utilities" Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960734 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e23875-5131-453e-a9d5-dcdbea0b4cf9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960741 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e23875-5131-453e-a9d5-dcdbea0b4cf9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960755 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="extract-content" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960760 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="extract-content" Nov 27 12:07:01 crc kubenswrapper[4796]: E1127 12:07:01.960774 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="registry-server" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960779 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="registry-server" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960952 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="98b2c708-f573-4d0d-b757-90217c0e633b" containerName="registry-server" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960975 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2bad3a-6372-484e-8107-2cec57c1b3a3" containerName="registry-server" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.960991 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e23875-5131-453e-a9d5-dcdbea0b4cf9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.961627 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.966641 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.966880 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.967023 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.972501 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:07:01 crc kubenswrapper[4796]: I1127 12:07:01.973930 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d"] Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.103754 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfh6q\" (UniqueName: \"kubernetes.io/projected/cb67680b-c997-4002-a238-67ff005b15f0-kube-api-access-vfh6q\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.103892 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.103995 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.205176 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.205245 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfh6q\" (UniqueName: \"kubernetes.io/projected/cb67680b-c997-4002-a238-67ff005b15f0-kube-api-access-vfh6q\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.205362 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.210255 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.211717 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.224340 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfh6q\" (UniqueName: \"kubernetes.io/projected/cb67680b-c997-4002-a238-67ff005b15f0-kube-api-access-vfh6q\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-krh4d\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.282135 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.791296 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d"] Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.801583 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:07:02 crc kubenswrapper[4796]: I1127 12:07:02.892967 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" event={"ID":"cb67680b-c997-4002-a238-67ff005b15f0","Type":"ContainerStarted","Data":"66988756b48ff7a78a840980054e1a2e9c05abc9250c8c0ea51777f664c1ae7b"} Nov 27 12:07:03 crc kubenswrapper[4796]: I1127 12:07:03.578985 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:07:03 crc kubenswrapper[4796]: E1127 12:07:03.579600 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:07:03 crc kubenswrapper[4796]: I1127 12:07:03.902396 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" event={"ID":"cb67680b-c997-4002-a238-67ff005b15f0","Type":"ContainerStarted","Data":"78234b6dc737eaec697a6b84fc7568ba8b4a8f7ee0cfe0fb6a5a2d31ba3ffaa2"} Nov 27 12:07:03 crc kubenswrapper[4796]: I1127 12:07:03.923902 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" podStartSLOduration=2.285601066 podStartE2EDuration="2.923877225s" podCreationTimestamp="2025-11-27 12:07:01 +0000 UTC" firstStartedPulling="2025-11-27 12:07:02.801195531 +0000 UTC m=+2540.319514449" lastFinishedPulling="2025-11-27 12:07:03.43947169 +0000 UTC m=+2540.957790608" observedRunningTime="2025-11-27 12:07:03.917386958 +0000 UTC m=+2541.435705876" watchObservedRunningTime="2025-11-27 12:07:03.923877225 +0000 UTC m=+2541.442196143" Nov 27 12:07:08 crc kubenswrapper[4796]: I1127 12:07:08.034434 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-276z6"] Nov 27 12:07:08 crc kubenswrapper[4796]: I1127 12:07:08.046534 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-276z6"] Nov 27 12:07:09 crc kubenswrapper[4796]: I1127 12:07:09.030392 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-km9nk"] Nov 27 12:07:09 crc kubenswrapper[4796]: I1127 12:07:09.042922 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-km9nk"] Nov 27 12:07:09 crc kubenswrapper[4796]: I1127 12:07:09.593176 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ffa4b0-e22e-4714-9b05-96bd3fffbfa8" path="/var/lib/kubelet/pods/23ffa4b0-e22e-4714-9b05-96bd3fffbfa8/volumes" Nov 27 12:07:09 crc kubenswrapper[4796]: I1127 12:07:09.595504 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d920f619-249d-4266-ae4b-e13917ae78dc" path="/var/lib/kubelet/pods/d920f619-249d-4266-ae4b-e13917ae78dc/volumes" Nov 27 12:07:10 crc kubenswrapper[4796]: I1127 12:07:10.031972 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-jfd4m"] Nov 27 12:07:10 crc kubenswrapper[4796]: I1127 12:07:10.041654 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-jfd4m"] Nov 27 12:07:11 crc kubenswrapper[4796]: I1127 12:07:11.579330 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d550fe-17da-4b68-8e62-b537d6c07314" path="/var/lib/kubelet/pods/24d550fe-17da-4b68-8e62-b537d6c07314/volumes" Nov 27 12:07:18 crc kubenswrapper[4796]: I1127 12:07:18.570583 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:07:18 crc kubenswrapper[4796]: E1127 12:07:18.574242 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:07:32 crc kubenswrapper[4796]: I1127 12:07:32.569213 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:07:32 crc kubenswrapper[4796]: E1127 12:07:32.569965 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:07:45 crc kubenswrapper[4796]: I1127 12:07:45.571309 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:07:45 crc kubenswrapper[4796]: E1127 12:07:45.572197 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:07:47 crc kubenswrapper[4796]: I1127 12:07:47.044048 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-67mjj"] Nov 27 12:07:47 crc kubenswrapper[4796]: I1127 12:07:47.054480 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-67mjj"] Nov 27 12:07:47 crc kubenswrapper[4796]: I1127 12:07:47.579287 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3" path="/var/lib/kubelet/pods/d3a975cf-a1d9-4cc3-ad76-4e60392c6fc3/volumes" Nov 27 12:07:48 crc kubenswrapper[4796]: I1127 12:07:48.044326 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5a51-account-create-update-s8rhc"] Nov 27 12:07:48 crc kubenswrapper[4796]: I1127 12:07:48.052833 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-jcs6h"] Nov 27 12:07:48 crc kubenswrapper[4796]: I1127 12:07:48.061151 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bbj62"] Nov 27 12:07:48 crc kubenswrapper[4796]: I1127 12:07:48.074017 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5a51-account-create-update-s8rhc"] Nov 27 12:07:48 crc kubenswrapper[4796]: I1127 12:07:48.082568 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bbj62"] Nov 27 12:07:48 crc kubenswrapper[4796]: I1127 12:07:48.090848 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-jcs6h"] Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.036833 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-4e96-account-create-update-dmgf7"] Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.047225 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e653-account-create-update-4wv78"] Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.055936 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-4e96-account-create-update-dmgf7"] Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.065301 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e653-account-create-update-4wv78"] Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.579897 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b498531-2aa8-4af7-bbc5-cfeb582a6813" path="/var/lib/kubelet/pods/0b498531-2aa8-4af7-bbc5-cfeb582a6813/volumes" Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.580799 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="813d270c-4a48-481f-ae6b-763f0f89fb43" path="/var/lib/kubelet/pods/813d270c-4a48-481f-ae6b-763f0f89fb43/volumes" Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.581345 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8af3632d-426e-4004-b787-66529242255b" path="/var/lib/kubelet/pods/8af3632d-426e-4004-b787-66529242255b/volumes" Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.581874 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcb76cd8-08df-45f1-ac77-6b8d326ed4d1" path="/var/lib/kubelet/pods/bcb76cd8-08df-45f1-ac77-6b8d326ed4d1/volumes" Nov 27 12:07:49 crc kubenswrapper[4796]: I1127 12:07:49.583033 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceebc602-7a6a-4fc6-baf7-483ca870a28f" path="/var/lib/kubelet/pods/ceebc602-7a6a-4fc6-baf7-483ca870a28f/volumes" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.331368 4796 scope.go:117] "RemoveContainer" containerID="44ea6ff19a93b29f6f3647cbb3c4692880f04249bbcfae1cd96727d2cb76b0b3" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.353839 4796 scope.go:117] "RemoveContainer" containerID="5794db6ea70b2cc1983c289178a1e8ff0817f382d265082865628ca8284b188f" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.403313 4796 scope.go:117] "RemoveContainer" containerID="33445dd4e42e13ddd1b17ea17ed7681779e513f0033ca2caa47a57e772114b46" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.464646 4796 scope.go:117] "RemoveContainer" containerID="c271eeeead555e12c731437950972e58c38d25b0fa872ef41340751282bcb26b" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.492679 4796 scope.go:117] "RemoveContainer" containerID="7d773520869c92070138018c552043b76571c30f0d46205320a726857974f3cc" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.537030 4796 scope.go:117] "RemoveContainer" containerID="8570a2265dd638c975fe57479b98f4c4297282d9a1715f2d868a1ef50ab1210b" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.581614 4796 scope.go:117] "RemoveContainer" containerID="2a542d127190aeed44c3c18627631a7b90870fa84e2b3f0cd41e93bc8514f27d" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.622782 4796 scope.go:117] "RemoveContainer" containerID="6ea459223ec6ce9b06b2dd9da5fcadb6836a5c10a7142318314ab8cf5707671b" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.642866 4796 scope.go:117] "RemoveContainer" containerID="e3e3526d04f419488f28c8c41e468a5086ac8c7dbaf8e74b572a62d692ec758b" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.665422 4796 scope.go:117] "RemoveContainer" containerID="7dc64cda3633c4e68703272a8f99797cb28b9dd31d7ef53257e67b9e59902190" Nov 27 12:07:51 crc kubenswrapper[4796]: I1127 12:07:51.693111 4796 scope.go:117] "RemoveContainer" containerID="d51dbf066d77c2ee6d11edcb605a311da42714d36fd4aff7c9918867e96fc70d" Nov 27 12:08:00 crc kubenswrapper[4796]: I1127 12:08:00.569903 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:08:00 crc kubenswrapper[4796]: E1127 12:08:00.570970 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:08:13 crc kubenswrapper[4796]: I1127 12:08:13.585872 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:08:13 crc kubenswrapper[4796]: E1127 12:08:13.588836 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:08:18 crc kubenswrapper[4796]: I1127 12:08:18.686606 4796 generic.go:334] "Generic (PLEG): container finished" podID="cb67680b-c997-4002-a238-67ff005b15f0" containerID="78234b6dc737eaec697a6b84fc7568ba8b4a8f7ee0cfe0fb6a5a2d31ba3ffaa2" exitCode=0 Nov 27 12:08:18 crc kubenswrapper[4796]: I1127 12:08:18.687204 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" event={"ID":"cb67680b-c997-4002-a238-67ff005b15f0","Type":"ContainerDied","Data":"78234b6dc737eaec697a6b84fc7568ba8b4a8f7ee0cfe0fb6a5a2d31ba3ffaa2"} Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.139628 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.217683 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-inventory\") pod \"cb67680b-c997-4002-a238-67ff005b15f0\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.217794 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfh6q\" (UniqueName: \"kubernetes.io/projected/cb67680b-c997-4002-a238-67ff005b15f0-kube-api-access-vfh6q\") pod \"cb67680b-c997-4002-a238-67ff005b15f0\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.217904 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-ssh-key\") pod \"cb67680b-c997-4002-a238-67ff005b15f0\" (UID: \"cb67680b-c997-4002-a238-67ff005b15f0\") " Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.223614 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb67680b-c997-4002-a238-67ff005b15f0-kube-api-access-vfh6q" (OuterVolumeSpecName: "kube-api-access-vfh6q") pod "cb67680b-c997-4002-a238-67ff005b15f0" (UID: "cb67680b-c997-4002-a238-67ff005b15f0"). InnerVolumeSpecName "kube-api-access-vfh6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.244818 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-inventory" (OuterVolumeSpecName: "inventory") pod "cb67680b-c997-4002-a238-67ff005b15f0" (UID: "cb67680b-c997-4002-a238-67ff005b15f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.245177 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cb67680b-c997-4002-a238-67ff005b15f0" (UID: "cb67680b-c997-4002-a238-67ff005b15f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.321384 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.321682 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfh6q\" (UniqueName: \"kubernetes.io/projected/cb67680b-c997-4002-a238-67ff005b15f0-kube-api-access-vfh6q\") on node \"crc\" DevicePath \"\"" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.321694 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb67680b-c997-4002-a238-67ff005b15f0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.704310 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" event={"ID":"cb67680b-c997-4002-a238-67ff005b15f0","Type":"ContainerDied","Data":"66988756b48ff7a78a840980054e1a2e9c05abc9250c8c0ea51777f664c1ae7b"} Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.704353 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-krh4d" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.704355 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66988756b48ff7a78a840980054e1a2e9c05abc9250c8c0ea51777f664c1ae7b" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.789632 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m"] Nov 27 12:08:20 crc kubenswrapper[4796]: E1127 12:08:20.790056 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb67680b-c997-4002-a238-67ff005b15f0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.790074 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb67680b-c997-4002-a238-67ff005b15f0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.790317 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb67680b-c997-4002-a238-67ff005b15f0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.791057 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.793788 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.795576 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.795670 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.795743 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.805969 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m"] Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.948785 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j2bd\" (UniqueName: \"kubernetes.io/projected/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-kube-api-access-4j2bd\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.949032 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:20 crc kubenswrapper[4796]: I1127 12:08:20.949354 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.051040 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.051151 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j2bd\" (UniqueName: \"kubernetes.io/projected/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-kube-api-access-4j2bd\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.051257 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.057329 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.067232 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.068513 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j2bd\" (UniqueName: \"kubernetes.io/projected/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-kube-api-access-4j2bd\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.154176 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.639581 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m"] Nov 27 12:08:21 crc kubenswrapper[4796]: I1127 12:08:21.729661 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" event={"ID":"8df61624-3a9b-4f6c-aa8a-2843e80fe64d","Type":"ContainerStarted","Data":"8addf6822b706659b6abf7dc716f3d41c67afad28c938aafe46b2ac689e695d1"} Nov 27 12:08:22 crc kubenswrapper[4796]: I1127 12:08:22.742521 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" event={"ID":"8df61624-3a9b-4f6c-aa8a-2843e80fe64d","Type":"ContainerStarted","Data":"207eaf5328deb61643ace92c99fdd9e291a7ed7d1597ff780f54164e7f682b8b"} Nov 27 12:08:22 crc kubenswrapper[4796]: I1127 12:08:22.765129 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" podStartSLOduration=2.026747101 podStartE2EDuration="2.765110984s" podCreationTimestamp="2025-11-27 12:08:20 +0000 UTC" firstStartedPulling="2025-11-27 12:08:21.646528232 +0000 UTC m=+2619.164847150" lastFinishedPulling="2025-11-27 12:08:22.384892105 +0000 UTC m=+2619.903211033" observedRunningTime="2025-11-27 12:08:22.756649273 +0000 UTC m=+2620.274968191" watchObservedRunningTime="2025-11-27 12:08:22.765110984 +0000 UTC m=+2620.283429902" Nov 27 12:08:24 crc kubenswrapper[4796]: I1127 12:08:24.569830 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:08:24 crc kubenswrapper[4796]: E1127 12:08:24.570549 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:08:25 crc kubenswrapper[4796]: I1127 12:08:25.037511 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-whbbq"] Nov 27 12:08:25 crc kubenswrapper[4796]: I1127 12:08:25.045975 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-whbbq"] Nov 27 12:08:25 crc kubenswrapper[4796]: I1127 12:08:25.582230 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6524424e-afc2-443c-beac-c0503ac10ddf" path="/var/lib/kubelet/pods/6524424e-afc2-443c-beac-c0503ac10ddf/volumes" Nov 27 12:08:27 crc kubenswrapper[4796]: I1127 12:08:27.807931 4796 generic.go:334] "Generic (PLEG): container finished" podID="8df61624-3a9b-4f6c-aa8a-2843e80fe64d" containerID="207eaf5328deb61643ace92c99fdd9e291a7ed7d1597ff780f54164e7f682b8b" exitCode=0 Nov 27 12:08:27 crc kubenswrapper[4796]: I1127 12:08:27.808038 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" event={"ID":"8df61624-3a9b-4f6c-aa8a-2843e80fe64d","Type":"ContainerDied","Data":"207eaf5328deb61643ace92c99fdd9e291a7ed7d1597ff780f54164e7f682b8b"} Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.298001 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.434893 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-inventory\") pod \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.435077 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-ssh-key\") pod \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.435229 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j2bd\" (UniqueName: \"kubernetes.io/projected/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-kube-api-access-4j2bd\") pod \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\" (UID: \"8df61624-3a9b-4f6c-aa8a-2843e80fe64d\") " Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.449222 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-kube-api-access-4j2bd" (OuterVolumeSpecName: "kube-api-access-4j2bd") pod "8df61624-3a9b-4f6c-aa8a-2843e80fe64d" (UID: "8df61624-3a9b-4f6c-aa8a-2843e80fe64d"). InnerVolumeSpecName "kube-api-access-4j2bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.463650 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-inventory" (OuterVolumeSpecName: "inventory") pod "8df61624-3a9b-4f6c-aa8a-2843e80fe64d" (UID: "8df61624-3a9b-4f6c-aa8a-2843e80fe64d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.464994 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8df61624-3a9b-4f6c-aa8a-2843e80fe64d" (UID: "8df61624-3a9b-4f6c-aa8a-2843e80fe64d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.537324 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.537357 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j2bd\" (UniqueName: \"kubernetes.io/projected/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-kube-api-access-4j2bd\") on node \"crc\" DevicePath \"\"" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.537368 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df61624-3a9b-4f6c-aa8a-2843e80fe64d-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.828130 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" event={"ID":"8df61624-3a9b-4f6c-aa8a-2843e80fe64d","Type":"ContainerDied","Data":"8addf6822b706659b6abf7dc716f3d41c67afad28c938aafe46b2ac689e695d1"} Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.828421 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8addf6822b706659b6abf7dc716f3d41c67afad28c938aafe46b2ac689e695d1" Nov 27 12:08:29 crc kubenswrapper[4796]: I1127 12:08:29.828379 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.007873 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7"] Nov 27 12:08:30 crc kubenswrapper[4796]: E1127 12:08:30.009532 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df61624-3a9b-4f6c-aa8a-2843e80fe64d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.009557 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df61624-3a9b-4f6c-aa8a-2843e80fe64d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.009768 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df61624-3a9b-4f6c-aa8a-2843e80fe64d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.010398 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.013206 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.013240 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.013828 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.013877 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.046099 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7"] Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.156638 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbfh5\" (UniqueName: \"kubernetes.io/projected/62906f9c-c8ce-48d2-b38b-db707e82f24c-kube-api-access-fbfh5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.156759 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.156786 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.259066 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbfh5\" (UniqueName: \"kubernetes.io/projected/62906f9c-c8ce-48d2-b38b-db707e82f24c-kube-api-access-fbfh5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.259173 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.259203 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.278525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.278674 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.281442 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbfh5\" (UniqueName: \"kubernetes.io/projected/62906f9c-c8ce-48d2-b38b-db707e82f24c-kube-api-access-fbfh5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bj5x7\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.331396 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:08:30 crc kubenswrapper[4796]: I1127 12:08:30.849662 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7"] Nov 27 12:08:31 crc kubenswrapper[4796]: I1127 12:08:31.851687 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" event={"ID":"62906f9c-c8ce-48d2-b38b-db707e82f24c","Type":"ContainerStarted","Data":"cb22d9d454d969b6e1f240c7ead4e6c0c4309ff128d257534241f6af2cc6f275"} Nov 27 12:08:32 crc kubenswrapper[4796]: I1127 12:08:32.864484 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" event={"ID":"62906f9c-c8ce-48d2-b38b-db707e82f24c","Type":"ContainerStarted","Data":"b2af6369eb4251fcda9156ca5efd19d3cc55aa20f2fda097a10cd02767c3b508"} Nov 27 12:08:32 crc kubenswrapper[4796]: I1127 12:08:32.884364 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" podStartSLOduration=3.123880865 podStartE2EDuration="3.884336013s" podCreationTimestamp="2025-11-27 12:08:29 +0000 UTC" firstStartedPulling="2025-11-27 12:08:30.856193509 +0000 UTC m=+2628.374512427" lastFinishedPulling="2025-11-27 12:08:31.616648657 +0000 UTC m=+2629.134967575" observedRunningTime="2025-11-27 12:08:32.878919444 +0000 UTC m=+2630.397238372" watchObservedRunningTime="2025-11-27 12:08:32.884336013 +0000 UTC m=+2630.402654931" Nov 27 12:08:36 crc kubenswrapper[4796]: I1127 12:08:36.569950 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:08:36 crc kubenswrapper[4796]: E1127 12:08:36.570457 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:08:44 crc kubenswrapper[4796]: I1127 12:08:44.033785 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lbg4j"] Nov 27 12:08:44 crc kubenswrapper[4796]: I1127 12:08:44.047243 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lbg4j"] Nov 27 12:08:45 crc kubenswrapper[4796]: I1127 12:08:45.587687 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cbdc050-92dd-44be-83f2-474365084f53" path="/var/lib/kubelet/pods/9cbdc050-92dd-44be-83f2-474365084f53/volumes" Nov 27 12:08:47 crc kubenswrapper[4796]: I1127 12:08:47.049297 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nzgfc"] Nov 27 12:08:47 crc kubenswrapper[4796]: I1127 12:08:47.056856 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nzgfc"] Nov 27 12:08:47 crc kubenswrapper[4796]: I1127 12:08:47.568695 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:08:47 crc kubenswrapper[4796]: E1127 12:08:47.569002 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:08:47 crc kubenswrapper[4796]: I1127 12:08:47.581158 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa3bdadd-98ed-4f1e-88c2-d23354a0e84a" path="/var/lib/kubelet/pods/aa3bdadd-98ed-4f1e-88c2-d23354a0e84a/volumes" Nov 27 12:08:51 crc kubenswrapper[4796]: I1127 12:08:51.898152 4796 scope.go:117] "RemoveContainer" containerID="5593861da04285f8c0fd914fe3b491e2e50de627ed3718f78d306a76e8e08b28" Nov 27 12:08:51 crc kubenswrapper[4796]: I1127 12:08:51.945633 4796 scope.go:117] "RemoveContainer" containerID="eac3bfb17e80603b35583d733348e26451d16279d1d6bd997f982c96b2f362a6" Nov 27 12:08:51 crc kubenswrapper[4796]: I1127 12:08:51.991935 4796 scope.go:117] "RemoveContainer" containerID="064006ab2a9695d2a447ea377d0daf5fc0c97676ebc78389fc16ebd4dd27c40d" Nov 27 12:08:59 crc kubenswrapper[4796]: I1127 12:08:59.569810 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:08:59 crc kubenswrapper[4796]: E1127 12:08:59.570539 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:09:10 crc kubenswrapper[4796]: I1127 12:09:10.216289 4796 generic.go:334] "Generic (PLEG): container finished" podID="62906f9c-c8ce-48d2-b38b-db707e82f24c" containerID="b2af6369eb4251fcda9156ca5efd19d3cc55aa20f2fda097a10cd02767c3b508" exitCode=0 Nov 27 12:09:10 crc kubenswrapper[4796]: I1127 12:09:10.216370 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" event={"ID":"62906f9c-c8ce-48d2-b38b-db707e82f24c","Type":"ContainerDied","Data":"b2af6369eb4251fcda9156ca5efd19d3cc55aa20f2fda097a10cd02767c3b508"} Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.640738 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.748452 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-inventory\") pod \"62906f9c-c8ce-48d2-b38b-db707e82f24c\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.748590 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-ssh-key\") pod \"62906f9c-c8ce-48d2-b38b-db707e82f24c\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.749504 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbfh5\" (UniqueName: \"kubernetes.io/projected/62906f9c-c8ce-48d2-b38b-db707e82f24c-kube-api-access-fbfh5\") pod \"62906f9c-c8ce-48d2-b38b-db707e82f24c\" (UID: \"62906f9c-c8ce-48d2-b38b-db707e82f24c\") " Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.754809 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62906f9c-c8ce-48d2-b38b-db707e82f24c-kube-api-access-fbfh5" (OuterVolumeSpecName: "kube-api-access-fbfh5") pod "62906f9c-c8ce-48d2-b38b-db707e82f24c" (UID: "62906f9c-c8ce-48d2-b38b-db707e82f24c"). InnerVolumeSpecName "kube-api-access-fbfh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.777626 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-inventory" (OuterVolumeSpecName: "inventory") pod "62906f9c-c8ce-48d2-b38b-db707e82f24c" (UID: "62906f9c-c8ce-48d2-b38b-db707e82f24c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.787578 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "62906f9c-c8ce-48d2-b38b-db707e82f24c" (UID: "62906f9c-c8ce-48d2-b38b-db707e82f24c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.852935 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbfh5\" (UniqueName: \"kubernetes.io/projected/62906f9c-c8ce-48d2-b38b-db707e82f24c-kube-api-access-fbfh5\") on node \"crc\" DevicePath \"\"" Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.852995 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:09:11 crc kubenswrapper[4796]: I1127 12:09:11.853009 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62906f9c-c8ce-48d2-b38b-db707e82f24c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.241978 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" event={"ID":"62906f9c-c8ce-48d2-b38b-db707e82f24c","Type":"ContainerDied","Data":"cb22d9d454d969b6e1f240c7ead4e6c0c4309ff128d257534241f6af2cc6f275"} Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.242017 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb22d9d454d969b6e1f240c7ead4e6c0c4309ff128d257534241f6af2cc6f275" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.242072 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bj5x7" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.327008 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44"] Nov 27 12:09:12 crc kubenswrapper[4796]: E1127 12:09:12.327941 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62906f9c-c8ce-48d2-b38b-db707e82f24c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.327963 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="62906f9c-c8ce-48d2-b38b-db707e82f24c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.328170 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="62906f9c-c8ce-48d2-b38b-db707e82f24c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.328810 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.330409 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.330503 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.330745 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.333473 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.344018 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44"] Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.464856 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.464945 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4hz9\" (UniqueName: \"kubernetes.io/projected/c41ab70c-1f9c-4961-9e0c-eea2a846f125-kube-api-access-q4hz9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.464971 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.567069 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.567181 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4hz9\" (UniqueName: \"kubernetes.io/projected/c41ab70c-1f9c-4961-9e0c-eea2a846f125-kube-api-access-q4hz9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.567227 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.573044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.581218 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.584692 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4hz9\" (UniqueName: \"kubernetes.io/projected/c41ab70c-1f9c-4961-9e0c-eea2a846f125-kube-api-access-q4hz9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kfn44\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:12 crc kubenswrapper[4796]: I1127 12:09:12.692088 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:09:13 crc kubenswrapper[4796]: I1127 12:09:13.232310 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44"] Nov 27 12:09:13 crc kubenswrapper[4796]: I1127 12:09:13.251006 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" event={"ID":"c41ab70c-1f9c-4961-9e0c-eea2a846f125","Type":"ContainerStarted","Data":"eda8b39e4e3438d2a59cc9afa09a00ff5d3e3547e231c2d75d47c90d865a54ca"} Nov 27 12:09:14 crc kubenswrapper[4796]: I1127 12:09:14.262025 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" event={"ID":"c41ab70c-1f9c-4961-9e0c-eea2a846f125","Type":"ContainerStarted","Data":"f3ff41f3c50ebfb1a81c79329a318ed353cdd63bf2828ae53a2c4bf73aa26574"} Nov 27 12:09:14 crc kubenswrapper[4796]: I1127 12:09:14.283219 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" podStartSLOduration=1.5911473470000002 podStartE2EDuration="2.283199155s" podCreationTimestamp="2025-11-27 12:09:12 +0000 UTC" firstStartedPulling="2025-11-27 12:09:13.238357958 +0000 UTC m=+2670.756676886" lastFinishedPulling="2025-11-27 12:09:13.930409776 +0000 UTC m=+2671.448728694" observedRunningTime="2025-11-27 12:09:14.280048119 +0000 UTC m=+2671.798367037" watchObservedRunningTime="2025-11-27 12:09:14.283199155 +0000 UTC m=+2671.801518073" Nov 27 12:09:14 crc kubenswrapper[4796]: I1127 12:09:14.568933 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:09:14 crc kubenswrapper[4796]: E1127 12:09:14.569569 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:09:26 crc kubenswrapper[4796]: I1127 12:09:26.569192 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:09:26 crc kubenswrapper[4796]: E1127 12:09:26.570058 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:09:30 crc kubenswrapper[4796]: I1127 12:09:30.075250 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-srjn9"] Nov 27 12:09:30 crc kubenswrapper[4796]: I1127 12:09:30.086673 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-srjn9"] Nov 27 12:09:31 crc kubenswrapper[4796]: I1127 12:09:31.588487 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d77f0e3-d2ec-485a-9616-a2fde63f185a" path="/var/lib/kubelet/pods/5d77f0e3-d2ec-485a-9616-a2fde63f185a/volumes" Nov 27 12:09:37 crc kubenswrapper[4796]: I1127 12:09:37.568998 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:09:37 crc kubenswrapper[4796]: E1127 12:09:37.570092 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:09:52 crc kubenswrapper[4796]: I1127 12:09:52.115731 4796 scope.go:117] "RemoveContainer" containerID="c500059c4dfad83633a876b9625956d648e9e712c9d714223b95c3a9dff4e27e" Nov 27 12:09:52 crc kubenswrapper[4796]: I1127 12:09:52.570638 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:09:52 crc kubenswrapper[4796]: E1127 12:09:52.571680 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:10:05 crc kubenswrapper[4796]: I1127 12:10:05.569287 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:10:05 crc kubenswrapper[4796]: E1127 12:10:05.570197 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:10:06 crc kubenswrapper[4796]: I1127 12:10:06.759619 4796 generic.go:334] "Generic (PLEG): container finished" podID="c41ab70c-1f9c-4961-9e0c-eea2a846f125" containerID="f3ff41f3c50ebfb1a81c79329a318ed353cdd63bf2828ae53a2c4bf73aa26574" exitCode=0 Nov 27 12:10:06 crc kubenswrapper[4796]: I1127 12:10:06.759715 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" event={"ID":"c41ab70c-1f9c-4961-9e0c-eea2a846f125","Type":"ContainerDied","Data":"f3ff41f3c50ebfb1a81c79329a318ed353cdd63bf2828ae53a2c4bf73aa26574"} Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.156203 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.347960 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4hz9\" (UniqueName: \"kubernetes.io/projected/c41ab70c-1f9c-4961-9e0c-eea2a846f125-kube-api-access-q4hz9\") pod \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.348137 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-inventory\") pod \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.348192 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-ssh-key\") pod \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\" (UID: \"c41ab70c-1f9c-4961-9e0c-eea2a846f125\") " Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.353694 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41ab70c-1f9c-4961-9e0c-eea2a846f125-kube-api-access-q4hz9" (OuterVolumeSpecName: "kube-api-access-q4hz9") pod "c41ab70c-1f9c-4961-9e0c-eea2a846f125" (UID: "c41ab70c-1f9c-4961-9e0c-eea2a846f125"). InnerVolumeSpecName "kube-api-access-q4hz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.386481 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-inventory" (OuterVolumeSpecName: "inventory") pod "c41ab70c-1f9c-4961-9e0c-eea2a846f125" (UID: "c41ab70c-1f9c-4961-9e0c-eea2a846f125"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.394516 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c41ab70c-1f9c-4961-9e0c-eea2a846f125" (UID: "c41ab70c-1f9c-4961-9e0c-eea2a846f125"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.450392 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4hz9\" (UniqueName: \"kubernetes.io/projected/c41ab70c-1f9c-4961-9e0c-eea2a846f125-kube-api-access-q4hz9\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.450439 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.450454 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c41ab70c-1f9c-4961-9e0c-eea2a846f125-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.779138 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" event={"ID":"c41ab70c-1f9c-4961-9e0c-eea2a846f125","Type":"ContainerDied","Data":"eda8b39e4e3438d2a59cc9afa09a00ff5d3e3547e231c2d75d47c90d865a54ca"} Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.779544 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eda8b39e4e3438d2a59cc9afa09a00ff5d3e3547e231c2d75d47c90d865a54ca" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.779205 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kfn44" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.929546 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ltbbg"] Nov 27 12:10:08 crc kubenswrapper[4796]: E1127 12:10:08.929960 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41ab70c-1f9c-4961-9e0c-eea2a846f125" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.929977 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41ab70c-1f9c-4961-9e0c-eea2a846f125" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.930198 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41ab70c-1f9c-4961-9e0c-eea2a846f125" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.930856 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.934384 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.934644 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.935001 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.935155 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:10:08 crc kubenswrapper[4796]: I1127 12:10:08.943458 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ltbbg"] Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.089260 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.089375 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpnw9\" (UniqueName: \"kubernetes.io/projected/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-kube-api-access-zpnw9\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.090107 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.192635 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpnw9\" (UniqueName: \"kubernetes.io/projected/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-kube-api-access-zpnw9\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.192737 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.192867 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.201943 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.201968 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.217785 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpnw9\" (UniqueName: \"kubernetes.io/projected/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-kube-api-access-zpnw9\") pod \"ssh-known-hosts-edpm-deployment-ltbbg\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.247817 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.580409 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ltbbg"] Nov 27 12:10:09 crc kubenswrapper[4796]: I1127 12:10:09.789997 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" event={"ID":"6b7f9b64-f1d9-4468-aafd-30819c86ab8a","Type":"ContainerStarted","Data":"cd7ff6b025e536a18260703bb72db433d06ec36da96a41008c4f10b6e26237b2"} Nov 27 12:10:10 crc kubenswrapper[4796]: I1127 12:10:10.799326 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" event={"ID":"6b7f9b64-f1d9-4468-aafd-30819c86ab8a","Type":"ContainerStarted","Data":"b7759ee0365346acfb5e3955639b4547565ca15d8740c5fde4a7784c5a7e0639"} Nov 27 12:10:10 crc kubenswrapper[4796]: I1127 12:10:10.817721 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" podStartSLOduration=2.1212027940000002 podStartE2EDuration="2.817700484s" podCreationTimestamp="2025-11-27 12:10:08 +0000 UTC" firstStartedPulling="2025-11-27 12:10:09.58521052 +0000 UTC m=+2727.103529428" lastFinishedPulling="2025-11-27 12:10:10.28170816 +0000 UTC m=+2727.800027118" observedRunningTime="2025-11-27 12:10:10.815437802 +0000 UTC m=+2728.333756720" watchObservedRunningTime="2025-11-27 12:10:10.817700484 +0000 UTC m=+2728.336019402" Nov 27 12:10:17 crc kubenswrapper[4796]: I1127 12:10:17.872251 4796 generic.go:334] "Generic (PLEG): container finished" podID="6b7f9b64-f1d9-4468-aafd-30819c86ab8a" containerID="b7759ee0365346acfb5e3955639b4547565ca15d8740c5fde4a7784c5a7e0639" exitCode=0 Nov 27 12:10:17 crc kubenswrapper[4796]: I1127 12:10:17.872377 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" event={"ID":"6b7f9b64-f1d9-4468-aafd-30819c86ab8a","Type":"ContainerDied","Data":"b7759ee0365346acfb5e3955639b4547565ca15d8740c5fde4a7784c5a7e0639"} Nov 27 12:10:18 crc kubenswrapper[4796]: I1127 12:10:18.569987 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:10:18 crc kubenswrapper[4796]: E1127 12:10:18.571102 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.288218 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.312916 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-inventory-0\") pod \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.313128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpnw9\" (UniqueName: \"kubernetes.io/projected/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-kube-api-access-zpnw9\") pod \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.313162 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-ssh-key-openstack-edpm-ipam\") pod \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\" (UID: \"6b7f9b64-f1d9-4468-aafd-30819c86ab8a\") " Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.320356 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-kube-api-access-zpnw9" (OuterVolumeSpecName: "kube-api-access-zpnw9") pod "6b7f9b64-f1d9-4468-aafd-30819c86ab8a" (UID: "6b7f9b64-f1d9-4468-aafd-30819c86ab8a"). InnerVolumeSpecName "kube-api-access-zpnw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.343419 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6b7f9b64-f1d9-4468-aafd-30819c86ab8a" (UID: "6b7f9b64-f1d9-4468-aafd-30819c86ab8a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.348516 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6b7f9b64-f1d9-4468-aafd-30819c86ab8a" (UID: "6b7f9b64-f1d9-4468-aafd-30819c86ab8a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.414998 4796 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.415032 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpnw9\" (UniqueName: \"kubernetes.io/projected/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-kube-api-access-zpnw9\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.415044 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6b7f9b64-f1d9-4468-aafd-30819c86ab8a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.893172 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" event={"ID":"6b7f9b64-f1d9-4468-aafd-30819c86ab8a","Type":"ContainerDied","Data":"cd7ff6b025e536a18260703bb72db433d06ec36da96a41008c4f10b6e26237b2"} Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.893213 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd7ff6b025e536a18260703bb72db433d06ec36da96a41008c4f10b6e26237b2" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.893308 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ltbbg" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.958121 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm"] Nov 27 12:10:19 crc kubenswrapper[4796]: E1127 12:10:19.958602 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7f9b64-f1d9-4468-aafd-30819c86ab8a" containerName="ssh-known-hosts-edpm-deployment" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.958623 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7f9b64-f1d9-4468-aafd-30819c86ab8a" containerName="ssh-known-hosts-edpm-deployment" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.958851 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7f9b64-f1d9-4468-aafd-30819c86ab8a" containerName="ssh-known-hosts-edpm-deployment" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.959642 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.961488 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.961544 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.961622 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.962032 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:10:19 crc kubenswrapper[4796]: I1127 12:10:19.984016 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm"] Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.027660 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chhpz\" (UniqueName: \"kubernetes.io/projected/eb6f1515-67d6-4070-ad15-fb26e89835d8-kube-api-access-chhpz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.027809 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.027997 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.130631 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chhpz\" (UniqueName: \"kubernetes.io/projected/eb6f1515-67d6-4070-ad15-fb26e89835d8-kube-api-access-chhpz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.130737 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.130819 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.136896 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.144729 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.147172 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chhpz\" (UniqueName: \"kubernetes.io/projected/eb6f1515-67d6-4070-ad15-fb26e89835d8-kube-api-access-chhpz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ldbkm\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.280886 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.800262 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm"] Nov 27 12:10:20 crc kubenswrapper[4796]: I1127 12:10:20.902467 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" event={"ID":"eb6f1515-67d6-4070-ad15-fb26e89835d8","Type":"ContainerStarted","Data":"645c0e32fd5b6ce93f186ba5c805076603828ed8dd8dbbc4e19e4de1cd8686bf"} Nov 27 12:10:21 crc kubenswrapper[4796]: I1127 12:10:21.929641 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" event={"ID":"eb6f1515-67d6-4070-ad15-fb26e89835d8","Type":"ContainerStarted","Data":"1487b8fcf583ccd1a40654586e32f2ed895f14ddb48cda1f7864e366b31f138f"} Nov 27 12:10:30 crc kubenswrapper[4796]: I1127 12:10:30.020953 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb6f1515-67d6-4070-ad15-fb26e89835d8" containerID="1487b8fcf583ccd1a40654586e32f2ed895f14ddb48cda1f7864e366b31f138f" exitCode=0 Nov 27 12:10:30 crc kubenswrapper[4796]: I1127 12:10:30.021030 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" event={"ID":"eb6f1515-67d6-4070-ad15-fb26e89835d8","Type":"ContainerDied","Data":"1487b8fcf583ccd1a40654586e32f2ed895f14ddb48cda1f7864e366b31f138f"} Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.472167 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.585252 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-ssh-key\") pod \"eb6f1515-67d6-4070-ad15-fb26e89835d8\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.586757 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-inventory\") pod \"eb6f1515-67d6-4070-ad15-fb26e89835d8\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.586839 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chhpz\" (UniqueName: \"kubernetes.io/projected/eb6f1515-67d6-4070-ad15-fb26e89835d8-kube-api-access-chhpz\") pod \"eb6f1515-67d6-4070-ad15-fb26e89835d8\" (UID: \"eb6f1515-67d6-4070-ad15-fb26e89835d8\") " Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.591477 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb6f1515-67d6-4070-ad15-fb26e89835d8-kube-api-access-chhpz" (OuterVolumeSpecName: "kube-api-access-chhpz") pod "eb6f1515-67d6-4070-ad15-fb26e89835d8" (UID: "eb6f1515-67d6-4070-ad15-fb26e89835d8"). InnerVolumeSpecName "kube-api-access-chhpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.617217 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-inventory" (OuterVolumeSpecName: "inventory") pod "eb6f1515-67d6-4070-ad15-fb26e89835d8" (UID: "eb6f1515-67d6-4070-ad15-fb26e89835d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.619015 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eb6f1515-67d6-4070-ad15-fb26e89835d8" (UID: "eb6f1515-67d6-4070-ad15-fb26e89835d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.689410 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.689441 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb6f1515-67d6-4070-ad15-fb26e89835d8-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:31 crc kubenswrapper[4796]: I1127 12:10:31.689450 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chhpz\" (UniqueName: \"kubernetes.io/projected/eb6f1515-67d6-4070-ad15-fb26e89835d8-kube-api-access-chhpz\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.039852 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" event={"ID":"eb6f1515-67d6-4070-ad15-fb26e89835d8","Type":"ContainerDied","Data":"645c0e32fd5b6ce93f186ba5c805076603828ed8dd8dbbc4e19e4de1cd8686bf"} Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.039891 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="645c0e32fd5b6ce93f186ba5c805076603828ed8dd8dbbc4e19e4de1cd8686bf" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.039952 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ldbkm" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.127335 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw"] Nov 27 12:10:32 crc kubenswrapper[4796]: E1127 12:10:32.127810 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb6f1515-67d6-4070-ad15-fb26e89835d8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.127830 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb6f1515-67d6-4070-ad15-fb26e89835d8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.128026 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb6f1515-67d6-4070-ad15-fb26e89835d8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.128704 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.131027 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.131308 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.131473 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.131706 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.139571 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw"] Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.196093 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pfrv\" (UniqueName: \"kubernetes.io/projected/bf61d6a5-49e7-4dc9-be44-0292579185c1-kube-api-access-6pfrv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.196496 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.196545 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.298049 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.298094 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.298125 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pfrv\" (UniqueName: \"kubernetes.io/projected/bf61d6a5-49e7-4dc9-be44-0292579185c1-kube-api-access-6pfrv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.302878 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.304414 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.318399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pfrv\" (UniqueName: \"kubernetes.io/projected/bf61d6a5-49e7-4dc9-be44-0292579185c1-kube-api-access-6pfrv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.457432 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:32 crc kubenswrapper[4796]: I1127 12:10:32.984116 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw"] Nov 27 12:10:33 crc kubenswrapper[4796]: I1127 12:10:33.049077 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" event={"ID":"bf61d6a5-49e7-4dc9-be44-0292579185c1","Type":"ContainerStarted","Data":"d28e414f13ebc813a3edcead60a228c52550e5b0f20b46c202c3762d1f0ecb0e"} Nov 27 12:10:33 crc kubenswrapper[4796]: I1127 12:10:33.589695 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:10:34 crc kubenswrapper[4796]: I1127 12:10:34.059163 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" event={"ID":"bf61d6a5-49e7-4dc9-be44-0292579185c1","Type":"ContainerStarted","Data":"8cf8c14124e32eeabd23acf152f6a7f2e97646e19d0ebfb7821e8fe0e7b2fd3f"} Nov 27 12:10:34 crc kubenswrapper[4796]: I1127 12:10:34.061617 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"5836e62a2bbad31fe21baa6664ea972f4b559db15b327f1ae4158570d6c719b1"} Nov 27 12:10:34 crc kubenswrapper[4796]: I1127 12:10:34.078844 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" podStartSLOduration=1.491426267 podStartE2EDuration="2.078825956s" podCreationTimestamp="2025-11-27 12:10:32 +0000 UTC" firstStartedPulling="2025-11-27 12:10:32.989081602 +0000 UTC m=+2750.507400520" lastFinishedPulling="2025-11-27 12:10:33.576481291 +0000 UTC m=+2751.094800209" observedRunningTime="2025-11-27 12:10:34.071300901 +0000 UTC m=+2751.589619839" watchObservedRunningTime="2025-11-27 12:10:34.078825956 +0000 UTC m=+2751.597144874" Nov 27 12:10:43 crc kubenswrapper[4796]: E1127 12:10:43.560378 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf61d6a5_49e7_4dc9_be44_0292579185c1.slice/crio-conmon-8cf8c14124e32eeabd23acf152f6a7f2e97646e19d0ebfb7821e8fe0e7b2fd3f.scope\": RecentStats: unable to find data in memory cache]" Nov 27 12:10:44 crc kubenswrapper[4796]: I1127 12:10:44.155110 4796 generic.go:334] "Generic (PLEG): container finished" podID="bf61d6a5-49e7-4dc9-be44-0292579185c1" containerID="8cf8c14124e32eeabd23acf152f6a7f2e97646e19d0ebfb7821e8fe0e7b2fd3f" exitCode=0 Nov 27 12:10:44 crc kubenswrapper[4796]: I1127 12:10:44.155319 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" event={"ID":"bf61d6a5-49e7-4dc9-be44-0292579185c1","Type":"ContainerDied","Data":"8cf8c14124e32eeabd23acf152f6a7f2e97646e19d0ebfb7821e8fe0e7b2fd3f"} Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.590625 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.774423 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-inventory\") pod \"bf61d6a5-49e7-4dc9-be44-0292579185c1\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.774519 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-ssh-key\") pod \"bf61d6a5-49e7-4dc9-be44-0292579185c1\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.774652 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pfrv\" (UniqueName: \"kubernetes.io/projected/bf61d6a5-49e7-4dc9-be44-0292579185c1-kube-api-access-6pfrv\") pod \"bf61d6a5-49e7-4dc9-be44-0292579185c1\" (UID: \"bf61d6a5-49e7-4dc9-be44-0292579185c1\") " Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.781477 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf61d6a5-49e7-4dc9-be44-0292579185c1-kube-api-access-6pfrv" (OuterVolumeSpecName: "kube-api-access-6pfrv") pod "bf61d6a5-49e7-4dc9-be44-0292579185c1" (UID: "bf61d6a5-49e7-4dc9-be44-0292579185c1"). InnerVolumeSpecName "kube-api-access-6pfrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.802553 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-inventory" (OuterVolumeSpecName: "inventory") pod "bf61d6a5-49e7-4dc9-be44-0292579185c1" (UID: "bf61d6a5-49e7-4dc9-be44-0292579185c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.832557 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf61d6a5-49e7-4dc9-be44-0292579185c1" (UID: "bf61d6a5-49e7-4dc9-be44-0292579185c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.876592 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.876628 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf61d6a5-49e7-4dc9-be44-0292579185c1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:45 crc kubenswrapper[4796]: I1127 12:10:45.876639 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pfrv\" (UniqueName: \"kubernetes.io/projected/bf61d6a5-49e7-4dc9-be44-0292579185c1-kube-api-access-6pfrv\") on node \"crc\" DevicePath \"\"" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.183763 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" event={"ID":"bf61d6a5-49e7-4dc9-be44-0292579185c1","Type":"ContainerDied","Data":"d28e414f13ebc813a3edcead60a228c52550e5b0f20b46c202c3762d1f0ecb0e"} Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.183807 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d28e414f13ebc813a3edcead60a228c52550e5b0f20b46c202c3762d1f0ecb0e" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.183869 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.261753 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb"] Nov 27 12:10:46 crc kubenswrapper[4796]: E1127 12:10:46.262223 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf61d6a5-49e7-4dc9-be44-0292579185c1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.262244 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf61d6a5-49e7-4dc9-be44-0292579185c1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.262563 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf61d6a5-49e7-4dc9-be44-0292579185c1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.263475 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.266600 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.266926 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.267185 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.267596 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.267689 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.267725 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.267755 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.268313 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.273696 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb"] Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389035 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqd48\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-kube-api-access-qqd48\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389107 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389136 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389189 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389212 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389371 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389429 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389466 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389512 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389594 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389693 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389733 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389774 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.389801 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.491986 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492041 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492087 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492116 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492229 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492311 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492348 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492390 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492459 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492519 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492569 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492631 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492664 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.492740 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqd48\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-kube-api-access-qqd48\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.498562 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.498802 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.499221 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.499291 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.499582 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.499812 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.500072 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.500379 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.500902 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.500985 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.501122 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.501175 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.502894 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.512124 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqd48\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-kube-api-access-qqd48\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:46 crc kubenswrapper[4796]: I1127 12:10:46.583320 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:10:47 crc kubenswrapper[4796]: I1127 12:10:47.126765 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb"] Nov 27 12:10:47 crc kubenswrapper[4796]: I1127 12:10:47.198181 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" event={"ID":"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9","Type":"ContainerStarted","Data":"ae39916647909a49285f11a4b05e23001dbe09ac52cc624ac063f33d03455936"} Nov 27 12:10:48 crc kubenswrapper[4796]: I1127 12:10:48.208399 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" event={"ID":"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9","Type":"ContainerStarted","Data":"df9c7f7dc5f6d177b1ac1f3e1355b5fc9f5de38dba463eb7e6a36ccf2cb630df"} Nov 27 12:10:48 crc kubenswrapper[4796]: I1127 12:10:48.227902 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" podStartSLOduration=1.777433509 podStartE2EDuration="2.227883197s" podCreationTimestamp="2025-11-27 12:10:46 +0000 UTC" firstStartedPulling="2025-11-27 12:10:47.137071414 +0000 UTC m=+2764.655390332" lastFinishedPulling="2025-11-27 12:10:47.587521062 +0000 UTC m=+2765.105840020" observedRunningTime="2025-11-27 12:10:48.223431186 +0000 UTC m=+2765.741750104" watchObservedRunningTime="2025-11-27 12:10:48.227883197 +0000 UTC m=+2765.746202115" Nov 27 12:11:26 crc kubenswrapper[4796]: I1127 12:11:26.526484 4796 generic.go:334] "Generic (PLEG): container finished" podID="77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" containerID="df9c7f7dc5f6d177b1ac1f3e1355b5fc9f5de38dba463eb7e6a36ccf2cb630df" exitCode=0 Nov 27 12:11:26 crc kubenswrapper[4796]: I1127 12:11:26.526606 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" event={"ID":"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9","Type":"ContainerDied","Data":"df9c7f7dc5f6d177b1ac1f3e1355b5fc9f5de38dba463eb7e6a36ccf2cb630df"} Nov 27 12:11:27 crc kubenswrapper[4796]: I1127 12:11:27.970980 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093413 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-telemetry-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093462 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-repo-setup-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093504 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqd48\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-kube-api-access-qqd48\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093542 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-bootstrap-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093619 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-inventory\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093641 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-nova-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093691 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ssh-key\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093735 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-neutron-metadata-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.093972 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.094015 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-libvirt-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.094034 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.094085 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.094110 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.094134 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ovn-combined-ca-bundle\") pod \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\" (UID: \"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9\") " Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.101837 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.102174 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.102194 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-kube-api-access-qqd48" (OuterVolumeSpecName: "kube-api-access-qqd48") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "kube-api-access-qqd48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.102475 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.103369 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.103553 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.104162 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.105241 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.106060 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.106111 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.107095 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.108614 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.132422 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.134985 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-inventory" (OuterVolumeSpecName: "inventory") pod "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" (UID: "77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210655 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqd48\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-kube-api-access-qqd48\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210693 4796 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210704 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210713 4796 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210722 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210732 4796 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210745 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210755 4796 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210764 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210772 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210782 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210791 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210801 4796 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.210810 4796 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.547307 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" event={"ID":"77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9","Type":"ContainerDied","Data":"ae39916647909a49285f11a4b05e23001dbe09ac52cc624ac063f33d03455936"} Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.547645 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae39916647909a49285f11a4b05e23001dbe09ac52cc624ac063f33d03455936" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.547701 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.646332 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l"] Nov 27 12:11:28 crc kubenswrapper[4796]: E1127 12:11:28.646784 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.646810 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.647133 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.653808 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.656358 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.656780 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.657072 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.657230 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.657427 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.662776 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l"] Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.719628 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.719706 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.719877 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.719936 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmsgx\" (UniqueName: \"kubernetes.io/projected/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-kube-api-access-cmsgx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.720062 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.822228 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.822344 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.822395 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.822459 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.822492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmsgx\" (UniqueName: \"kubernetes.io/projected/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-kube-api-access-cmsgx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.823566 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.826797 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.827884 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.835925 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:28 crc kubenswrapper[4796]: I1127 12:11:28.838990 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmsgx\" (UniqueName: \"kubernetes.io/projected/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-kube-api-access-cmsgx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7gg7l\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:29 crc kubenswrapper[4796]: I1127 12:11:29.029943 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:11:29 crc kubenswrapper[4796]: I1127 12:11:29.548588 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l"] Nov 27 12:11:30 crc kubenswrapper[4796]: I1127 12:11:30.564502 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" event={"ID":"4b5e628e-9f89-4ae6-80c0-74d9cf57535d","Type":"ContainerStarted","Data":"f595d3e494d4f5cfd67015fa06d5d849ab48357c578fec43f331b181d0816632"} Nov 27 12:11:31 crc kubenswrapper[4796]: I1127 12:11:31.582117 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" event={"ID":"4b5e628e-9f89-4ae6-80c0-74d9cf57535d","Type":"ContainerStarted","Data":"3ccaa9f0e6df85457998f651ad2b30839d874fa7b6c896d14595afc1f33c3b16"} Nov 27 12:11:31 crc kubenswrapper[4796]: I1127 12:11:31.612747 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" podStartSLOduration=2.534394235 podStartE2EDuration="3.612728106s" podCreationTimestamp="2025-11-27 12:11:28 +0000 UTC" firstStartedPulling="2025-11-27 12:11:29.554412039 +0000 UTC m=+2807.072730957" lastFinishedPulling="2025-11-27 12:11:30.6327459 +0000 UTC m=+2808.151064828" observedRunningTime="2025-11-27 12:11:31.607566687 +0000 UTC m=+2809.125885625" watchObservedRunningTime="2025-11-27 12:11:31.612728106 +0000 UTC m=+2809.131047024" Nov 27 12:12:15 crc kubenswrapper[4796]: I1127 12:12:15.615880 4796 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6568564dcc-mst5g" podUID="c65b890a-db52-4342-938c-d56538e76b3c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 27 12:12:34 crc kubenswrapper[4796]: I1127 12:12:34.153585 4796 generic.go:334] "Generic (PLEG): container finished" podID="4b5e628e-9f89-4ae6-80c0-74d9cf57535d" containerID="3ccaa9f0e6df85457998f651ad2b30839d874fa7b6c896d14595afc1f33c3b16" exitCode=0 Nov 27 12:12:34 crc kubenswrapper[4796]: I1127 12:12:34.153669 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" event={"ID":"4b5e628e-9f89-4ae6-80c0-74d9cf57535d","Type":"ContainerDied","Data":"3ccaa9f0e6df85457998f651ad2b30839d874fa7b6c896d14595afc1f33c3b16"} Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.586519 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.657800 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmsgx\" (UniqueName: \"kubernetes.io/projected/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-kube-api-access-cmsgx\") pod \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.658031 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovncontroller-config-0\") pod \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.658128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ssh-key\") pod \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.658211 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovn-combined-ca-bundle\") pod \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.658303 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-inventory\") pod \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\" (UID: \"4b5e628e-9f89-4ae6-80c0-74d9cf57535d\") " Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.663024 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4b5e628e-9f89-4ae6-80c0-74d9cf57535d" (UID: "4b5e628e-9f89-4ae6-80c0-74d9cf57535d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.663898 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-kube-api-access-cmsgx" (OuterVolumeSpecName: "kube-api-access-cmsgx") pod "4b5e628e-9f89-4ae6-80c0-74d9cf57535d" (UID: "4b5e628e-9f89-4ae6-80c0-74d9cf57535d"). InnerVolumeSpecName "kube-api-access-cmsgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.684289 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-inventory" (OuterVolumeSpecName: "inventory") pod "4b5e628e-9f89-4ae6-80c0-74d9cf57535d" (UID: "4b5e628e-9f89-4ae6-80c0-74d9cf57535d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.687774 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4b5e628e-9f89-4ae6-80c0-74d9cf57535d" (UID: "4b5e628e-9f89-4ae6-80c0-74d9cf57535d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.695173 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b5e628e-9f89-4ae6-80c0-74d9cf57535d" (UID: "4b5e628e-9f89-4ae6-80c0-74d9cf57535d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.761123 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmsgx\" (UniqueName: \"kubernetes.io/projected/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-kube-api-access-cmsgx\") on node \"crc\" DevicePath \"\"" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.761508 4796 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.761619 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.761723 4796 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:12:35 crc kubenswrapper[4796]: I1127 12:12:35.761824 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b5e628e-9f89-4ae6-80c0-74d9cf57535d-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.177157 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" event={"ID":"4b5e628e-9f89-4ae6-80c0-74d9cf57535d","Type":"ContainerDied","Data":"f595d3e494d4f5cfd67015fa06d5d849ab48357c578fec43f331b181d0816632"} Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.177537 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f595d3e494d4f5cfd67015fa06d5d849ab48357c578fec43f331b181d0816632" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.177223 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7gg7l" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.264753 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z"] Nov 27 12:12:36 crc kubenswrapper[4796]: E1127 12:12:36.265553 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b5e628e-9f89-4ae6-80c0-74d9cf57535d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.265570 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b5e628e-9f89-4ae6-80c0-74d9cf57535d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.266064 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b5e628e-9f89-4ae6-80c0-74d9cf57535d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.267081 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.273217 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.273282 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.273350 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.273401 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.273436 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfwcx\" (UniqueName: \"kubernetes.io/projected/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-kube-api-access-wfwcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.273534 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.311707 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.311742 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.312136 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.312248 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z"] Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.312577 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.312588 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.312841 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.375222 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.375349 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.375381 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfwcx\" (UniqueName: \"kubernetes.io/projected/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-kube-api-access-wfwcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.375486 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.375572 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.375599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.379946 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.380351 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.380741 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.381108 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.392388 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfwcx\" (UniqueName: \"kubernetes.io/projected/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-kube-api-access-wfwcx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: E1127 12:12:36.392385 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b5e628e_9f89_4ae6_80c0_74d9cf57535d.slice\": RecentStats: unable to find data in memory cache]" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.398040 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:36 crc kubenswrapper[4796]: I1127 12:12:36.635087 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:12:37 crc kubenswrapper[4796]: I1127 12:12:37.152298 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z"] Nov 27 12:12:37 crc kubenswrapper[4796]: W1127 12:12:37.162691 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod213166c4_bd90_4e0f_8fbb_1164ebc31cb8.slice/crio-e445ac5a126dc82784f3344c49b1bce169418d31c56124f2c63cf1f852f9075d WatchSource:0}: Error finding container e445ac5a126dc82784f3344c49b1bce169418d31c56124f2c63cf1f852f9075d: Status 404 returned error can't find the container with id e445ac5a126dc82784f3344c49b1bce169418d31c56124f2c63cf1f852f9075d Nov 27 12:12:37 crc kubenswrapper[4796]: I1127 12:12:37.165353 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:12:37 crc kubenswrapper[4796]: I1127 12:12:37.190148 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" event={"ID":"213166c4-bd90-4e0f-8fbb-1164ebc31cb8","Type":"ContainerStarted","Data":"e445ac5a126dc82784f3344c49b1bce169418d31c56124f2c63cf1f852f9075d"} Nov 27 12:12:38 crc kubenswrapper[4796]: I1127 12:12:38.202170 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" event={"ID":"213166c4-bd90-4e0f-8fbb-1164ebc31cb8","Type":"ContainerStarted","Data":"2f49e85428283dd0e7d94eb18d5748b4df2ed9ec1825704fcd665f70da427763"} Nov 27 12:12:38 crc kubenswrapper[4796]: I1127 12:12:38.225666 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" podStartSLOduration=1.556079971 podStartE2EDuration="2.22564841s" podCreationTimestamp="2025-11-27 12:12:36 +0000 UTC" firstStartedPulling="2025-11-27 12:12:37.165076389 +0000 UTC m=+2874.683395307" lastFinishedPulling="2025-11-27 12:12:37.834644838 +0000 UTC m=+2875.352963746" observedRunningTime="2025-11-27 12:12:38.218025374 +0000 UTC m=+2875.736344292" watchObservedRunningTime="2025-11-27 12:12:38.22564841 +0000 UTC m=+2875.743967328" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.036083 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kx9f7"] Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.040254 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.048928 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kx9f7"] Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.107177 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-catalog-content\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.107228 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58zhl\" (UniqueName: \"kubernetes.io/projected/dbea2432-846a-4574-9fc4-87898c53a8e6-kube-api-access-58zhl\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.107687 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-utilities\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.209337 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-utilities\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.209429 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-catalog-content\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.209472 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58zhl\" (UniqueName: \"kubernetes.io/projected/dbea2432-846a-4574-9fc4-87898c53a8e6-kube-api-access-58zhl\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.209906 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-utilities\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.209943 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-catalog-content\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.228886 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58zhl\" (UniqueName: \"kubernetes.io/projected/dbea2432-846a-4574-9fc4-87898c53a8e6-kube-api-access-58zhl\") pod \"certified-operators-kx9f7\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.372218 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:47 crc kubenswrapper[4796]: I1127 12:12:47.904701 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kx9f7"] Nov 27 12:12:48 crc kubenswrapper[4796]: I1127 12:12:48.304187 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerID="fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145" exitCode=0 Nov 27 12:12:48 crc kubenswrapper[4796]: I1127 12:12:48.304231 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerDied","Data":"fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145"} Nov 27 12:12:48 crc kubenswrapper[4796]: I1127 12:12:48.304505 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerStarted","Data":"a54c799fbdb9fa6ce74209418632827254f3d33d5e7fb9e297af639e20d461c7"} Nov 27 12:12:49 crc kubenswrapper[4796]: I1127 12:12:49.315082 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerStarted","Data":"2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301"} Nov 27 12:12:50 crc kubenswrapper[4796]: I1127 12:12:50.329418 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerID="2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301" exitCode=0 Nov 27 12:12:50 crc kubenswrapper[4796]: I1127 12:12:50.329475 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerDied","Data":"2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301"} Nov 27 12:12:51 crc kubenswrapper[4796]: I1127 12:12:51.340254 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerStarted","Data":"04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c"} Nov 27 12:12:51 crc kubenswrapper[4796]: I1127 12:12:51.356449 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kx9f7" podStartSLOduration=1.5468143699999999 podStartE2EDuration="4.356429372s" podCreationTimestamp="2025-11-27 12:12:47 +0000 UTC" firstStartedPulling="2025-11-27 12:12:48.30631495 +0000 UTC m=+2885.824633868" lastFinishedPulling="2025-11-27 12:12:51.115929952 +0000 UTC m=+2888.634248870" observedRunningTime="2025-11-27 12:12:51.35635542 +0000 UTC m=+2888.874674338" watchObservedRunningTime="2025-11-27 12:12:51.356429372 +0000 UTC m=+2888.874748320" Nov 27 12:12:57 crc kubenswrapper[4796]: I1127 12:12:57.373157 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:57 crc kubenswrapper[4796]: I1127 12:12:57.373755 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:57 crc kubenswrapper[4796]: I1127 12:12:57.427825 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:57 crc kubenswrapper[4796]: I1127 12:12:57.477812 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:12:57 crc kubenswrapper[4796]: I1127 12:12:57.665496 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kx9f7"] Nov 27 12:12:59 crc kubenswrapper[4796]: I1127 12:12:59.409639 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kx9f7" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="registry-server" containerID="cri-o://04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c" gracePeriod=2 Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.204201 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.284479 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58zhl\" (UniqueName: \"kubernetes.io/projected/dbea2432-846a-4574-9fc4-87898c53a8e6-kube-api-access-58zhl\") pod \"dbea2432-846a-4574-9fc4-87898c53a8e6\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.284855 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-utilities\") pod \"dbea2432-846a-4574-9fc4-87898c53a8e6\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.285064 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-catalog-content\") pod \"dbea2432-846a-4574-9fc4-87898c53a8e6\" (UID: \"dbea2432-846a-4574-9fc4-87898c53a8e6\") " Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.285855 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-utilities" (OuterVolumeSpecName: "utilities") pod "dbea2432-846a-4574-9fc4-87898c53a8e6" (UID: "dbea2432-846a-4574-9fc4-87898c53a8e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.291082 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbea2432-846a-4574-9fc4-87898c53a8e6-kube-api-access-58zhl" (OuterVolumeSpecName: "kube-api-access-58zhl") pod "dbea2432-846a-4574-9fc4-87898c53a8e6" (UID: "dbea2432-846a-4574-9fc4-87898c53a8e6"). InnerVolumeSpecName "kube-api-access-58zhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.342132 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbea2432-846a-4574-9fc4-87898c53a8e6" (UID: "dbea2432-846a-4574-9fc4-87898c53a8e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.387557 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.387596 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbea2432-846a-4574-9fc4-87898c53a8e6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.387611 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58zhl\" (UniqueName: \"kubernetes.io/projected/dbea2432-846a-4574-9fc4-87898c53a8e6-kube-api-access-58zhl\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.421442 4796 generic.go:334] "Generic (PLEG): container finished" podID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerID="04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c" exitCode=0 Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.421522 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kx9f7" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.421541 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerDied","Data":"04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c"} Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.422662 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kx9f7" event={"ID":"dbea2432-846a-4574-9fc4-87898c53a8e6","Type":"ContainerDied","Data":"a54c799fbdb9fa6ce74209418632827254f3d33d5e7fb9e297af639e20d461c7"} Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.422688 4796 scope.go:117] "RemoveContainer" containerID="04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.441171 4796 scope.go:117] "RemoveContainer" containerID="2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.463833 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kx9f7"] Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.471928 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kx9f7"] Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.474255 4796 scope.go:117] "RemoveContainer" containerID="fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.515240 4796 scope.go:117] "RemoveContainer" containerID="04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c" Nov 27 12:13:00 crc kubenswrapper[4796]: E1127 12:13:00.515665 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c\": container with ID starting with 04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c not found: ID does not exist" containerID="04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.515700 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c"} err="failed to get container status \"04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c\": rpc error: code = NotFound desc = could not find container \"04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c\": container with ID starting with 04a9af624a9d7992acd22c28cc11dd22fca08f6c81aee23fbfbf25ea0c02997c not found: ID does not exist" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.515723 4796 scope.go:117] "RemoveContainer" containerID="2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301" Nov 27 12:13:00 crc kubenswrapper[4796]: E1127 12:13:00.515937 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301\": container with ID starting with 2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301 not found: ID does not exist" containerID="2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.515955 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301"} err="failed to get container status \"2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301\": rpc error: code = NotFound desc = could not find container \"2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301\": container with ID starting with 2a3a3f1aa9a69581642d2c1cd0a8060fd78874473866f1c1e5ad883fc865f301 not found: ID does not exist" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.515967 4796 scope.go:117] "RemoveContainer" containerID="fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145" Nov 27 12:13:00 crc kubenswrapper[4796]: E1127 12:13:00.516167 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145\": container with ID starting with fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145 not found: ID does not exist" containerID="fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145" Nov 27 12:13:00 crc kubenswrapper[4796]: I1127 12:13:00.516191 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145"} err="failed to get container status \"fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145\": rpc error: code = NotFound desc = could not find container \"fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145\": container with ID starting with fbd225d7f9eb1dbee965d7b6c16dcd8e6a401b3897973de9d20e2919a82f1145 not found: ID does not exist" Nov 27 12:13:01 crc kubenswrapper[4796]: I1127 12:13:01.583788 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" path="/var/lib/kubelet/pods/dbea2432-846a-4574-9fc4-87898c53a8e6/volumes" Nov 27 12:13:01 crc kubenswrapper[4796]: I1127 12:13:01.879988 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:13:01 crc kubenswrapper[4796]: I1127 12:13:01.880049 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:13:26 crc kubenswrapper[4796]: I1127 12:13:26.666859 4796 generic.go:334] "Generic (PLEG): container finished" podID="213166c4-bd90-4e0f-8fbb-1164ebc31cb8" containerID="2f49e85428283dd0e7d94eb18d5748b4df2ed9ec1825704fcd665f70da427763" exitCode=0 Nov 27 12:13:26 crc kubenswrapper[4796]: I1127 12:13:26.666928 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" event={"ID":"213166c4-bd90-4e0f-8fbb-1164ebc31cb8","Type":"ContainerDied","Data":"2f49e85428283dd0e7d94eb18d5748b4df2ed9ec1825704fcd665f70da427763"} Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.167124 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.230457 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-inventory\") pod \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.230503 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-ssh-key\") pod \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.230529 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-metadata-combined-ca-bundle\") pod \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.230649 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfwcx\" (UniqueName: \"kubernetes.io/projected/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-kube-api-access-wfwcx\") pod \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.230700 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.230745 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-nova-metadata-neutron-config-0\") pod \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\" (UID: \"213166c4-bd90-4e0f-8fbb-1164ebc31cb8\") " Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.236566 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "213166c4-bd90-4e0f-8fbb-1164ebc31cb8" (UID: "213166c4-bd90-4e0f-8fbb-1164ebc31cb8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.237495 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-kube-api-access-wfwcx" (OuterVolumeSpecName: "kube-api-access-wfwcx") pod "213166c4-bd90-4e0f-8fbb-1164ebc31cb8" (UID: "213166c4-bd90-4e0f-8fbb-1164ebc31cb8"). InnerVolumeSpecName "kube-api-access-wfwcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.259465 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "213166c4-bd90-4e0f-8fbb-1164ebc31cb8" (UID: "213166c4-bd90-4e0f-8fbb-1164ebc31cb8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.265683 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "213166c4-bd90-4e0f-8fbb-1164ebc31cb8" (UID: "213166c4-bd90-4e0f-8fbb-1164ebc31cb8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.266770 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "213166c4-bd90-4e0f-8fbb-1164ebc31cb8" (UID: "213166c4-bd90-4e0f-8fbb-1164ebc31cb8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.269539 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-inventory" (OuterVolumeSpecName: "inventory") pod "213166c4-bd90-4e0f-8fbb-1164ebc31cb8" (UID: "213166c4-bd90-4e0f-8fbb-1164ebc31cb8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.333340 4796 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.333376 4796 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.333386 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.333395 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.333404 4796 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.333413 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfwcx\" (UniqueName: \"kubernetes.io/projected/213166c4-bd90-4e0f-8fbb-1164ebc31cb8-kube-api-access-wfwcx\") on node \"crc\" DevicePath \"\"" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.689851 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" event={"ID":"213166c4-bd90-4e0f-8fbb-1164ebc31cb8","Type":"ContainerDied","Data":"e445ac5a126dc82784f3344c49b1bce169418d31c56124f2c63cf1f852f9075d"} Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.690302 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e445ac5a126dc82784f3344c49b1bce169418d31c56124f2c63cf1f852f9075d" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.689977 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.832931 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8"] Nov 27 12:13:28 crc kubenswrapper[4796]: E1127 12:13:28.833450 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="extract-content" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.833472 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="extract-content" Nov 27 12:13:28 crc kubenswrapper[4796]: E1127 12:13:28.833483 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="registry-server" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.833492 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="registry-server" Nov 27 12:13:28 crc kubenswrapper[4796]: E1127 12:13:28.833501 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="extract-utilities" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.833507 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="extract-utilities" Nov 27 12:13:28 crc kubenswrapper[4796]: E1127 12:13:28.833541 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="213166c4-bd90-4e0f-8fbb-1164ebc31cb8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.833549 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="213166c4-bd90-4e0f-8fbb-1164ebc31cb8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.833739 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbea2432-846a-4574-9fc4-87898c53a8e6" containerName="registry-server" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.833759 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="213166c4-bd90-4e0f-8fbb-1164ebc31cb8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.834571 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.836865 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.837176 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.837865 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.838112 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.838813 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.839420 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.839461 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjqp5\" (UniqueName: \"kubernetes.io/projected/a779ff89-a04b-41a9-acfe-64e318bbf745-kube-api-access-wjqp5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.839691 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.839854 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.839959 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.857536 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8"] Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.941492 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.941558 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjqp5\" (UniqueName: \"kubernetes.io/projected/a779ff89-a04b-41a9-acfe-64e318bbf745-kube-api-access-wjqp5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.941608 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.941653 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.941687 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.946383 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.946618 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.947830 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.955757 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:28 crc kubenswrapper[4796]: I1127 12:13:28.963569 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjqp5\" (UniqueName: \"kubernetes.io/projected/a779ff89-a04b-41a9-acfe-64e318bbf745-kube-api-access-wjqp5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:29 crc kubenswrapper[4796]: I1127 12:13:29.155681 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:13:29 crc kubenswrapper[4796]: I1127 12:13:29.639232 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8"] Nov 27 12:13:29 crc kubenswrapper[4796]: I1127 12:13:29.698869 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" event={"ID":"a779ff89-a04b-41a9-acfe-64e318bbf745","Type":"ContainerStarted","Data":"75657f5fa5bb8a36703f9518f2719d8deb655315efb80ce97353e3d4113a9818"} Nov 27 12:13:31 crc kubenswrapper[4796]: I1127 12:13:31.726945 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" event={"ID":"a779ff89-a04b-41a9-acfe-64e318bbf745","Type":"ContainerStarted","Data":"29a76dbd19f3d551770239a98c53d1492336de56369116ec78791b17d003112a"} Nov 27 12:13:31 crc kubenswrapper[4796]: I1127 12:13:31.748211 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" podStartSLOduration=2.774131025 podStartE2EDuration="3.748193751s" podCreationTimestamp="2025-11-27 12:13:28 +0000 UTC" firstStartedPulling="2025-11-27 12:13:29.650922243 +0000 UTC m=+2927.169241161" lastFinishedPulling="2025-11-27 12:13:30.624984959 +0000 UTC m=+2928.143303887" observedRunningTime="2025-11-27 12:13:31.747192474 +0000 UTC m=+2929.265511412" watchObservedRunningTime="2025-11-27 12:13:31.748193751 +0000 UTC m=+2929.266512679" Nov 27 12:13:31 crc kubenswrapper[4796]: I1127 12:13:31.879670 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:13:31 crc kubenswrapper[4796]: I1127 12:13:31.879726 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:14:01 crc kubenswrapper[4796]: I1127 12:14:01.879947 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:14:01 crc kubenswrapper[4796]: I1127 12:14:01.880675 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:14:01 crc kubenswrapper[4796]: I1127 12:14:01.880739 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:14:01 crc kubenswrapper[4796]: I1127 12:14:01.881557 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5836e62a2bbad31fe21baa6664ea972f4b559db15b327f1ae4158570d6c719b1"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:14:01 crc kubenswrapper[4796]: I1127 12:14:01.881620 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://5836e62a2bbad31fe21baa6664ea972f4b559db15b327f1ae4158570d6c719b1" gracePeriod=600 Nov 27 12:14:03 crc kubenswrapper[4796]: I1127 12:14:03.002227 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="5836e62a2bbad31fe21baa6664ea972f4b559db15b327f1ae4158570d6c719b1" exitCode=0 Nov 27 12:14:03 crc kubenswrapper[4796]: I1127 12:14:03.002292 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"5836e62a2bbad31fe21baa6664ea972f4b559db15b327f1ae4158570d6c719b1"} Nov 27 12:14:03 crc kubenswrapper[4796]: I1127 12:14:03.002892 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e"} Nov 27 12:14:03 crc kubenswrapper[4796]: I1127 12:14:03.002916 4796 scope.go:117] "RemoveContainer" containerID="fe26a9d55dcfbded04fe607652b77b828c7ce0c8604675c5a33596050c34f32f" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.170096 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c"] Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.172018 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.174993 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.175554 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.188742 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c"] Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.249194 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2zvb\" (UniqueName: \"kubernetes.io/projected/20af4cb7-d594-4ea4-ad49-ddbf806557a2-kube-api-access-d2zvb\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.249559 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20af4cb7-d594-4ea4-ad49-ddbf806557a2-secret-volume\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.249790 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20af4cb7-d594-4ea4-ad49-ddbf806557a2-config-volume\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.351340 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2zvb\" (UniqueName: \"kubernetes.io/projected/20af4cb7-d594-4ea4-ad49-ddbf806557a2-kube-api-access-d2zvb\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.351749 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20af4cb7-d594-4ea4-ad49-ddbf806557a2-secret-volume\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.351857 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20af4cb7-d594-4ea4-ad49-ddbf806557a2-config-volume\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.352932 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20af4cb7-d594-4ea4-ad49-ddbf806557a2-config-volume\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.358047 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20af4cb7-d594-4ea4-ad49-ddbf806557a2-secret-volume\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.370042 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2zvb\" (UniqueName: \"kubernetes.io/projected/20af4cb7-d594-4ea4-ad49-ddbf806557a2-kube-api-access-d2zvb\") pod \"collect-profiles-29404095-dvr6c\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.509702 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:00 crc kubenswrapper[4796]: I1127 12:15:00.947193 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c"] Nov 27 12:15:01 crc kubenswrapper[4796]: I1127 12:15:01.513755 4796 generic.go:334] "Generic (PLEG): container finished" podID="20af4cb7-d594-4ea4-ad49-ddbf806557a2" containerID="0ff8ac8889df4943c1942e96adedaca4e17df8aea13f60fe25087606a55cd099" exitCode=0 Nov 27 12:15:01 crc kubenswrapper[4796]: I1127 12:15:01.513829 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" event={"ID":"20af4cb7-d594-4ea4-ad49-ddbf806557a2","Type":"ContainerDied","Data":"0ff8ac8889df4943c1942e96adedaca4e17df8aea13f60fe25087606a55cd099"} Nov 27 12:15:01 crc kubenswrapper[4796]: I1127 12:15:01.514099 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" event={"ID":"20af4cb7-d594-4ea4-ad49-ddbf806557a2","Type":"ContainerStarted","Data":"766f7d481db074b2d5cb71e512b9b3f78040434b7a54acad1bd416d1b9820563"} Nov 27 12:15:02 crc kubenswrapper[4796]: I1127 12:15:02.858071 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:02 crc kubenswrapper[4796]: I1127 12:15:02.997863 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20af4cb7-d594-4ea4-ad49-ddbf806557a2-secret-volume\") pod \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " Nov 27 12:15:02 crc kubenswrapper[4796]: I1127 12:15:02.997943 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2zvb\" (UniqueName: \"kubernetes.io/projected/20af4cb7-d594-4ea4-ad49-ddbf806557a2-kube-api-access-d2zvb\") pod \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " Nov 27 12:15:02 crc kubenswrapper[4796]: I1127 12:15:02.997995 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20af4cb7-d594-4ea4-ad49-ddbf806557a2-config-volume\") pod \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\" (UID: \"20af4cb7-d594-4ea4-ad49-ddbf806557a2\") " Nov 27 12:15:02 crc kubenswrapper[4796]: I1127 12:15:02.998913 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20af4cb7-d594-4ea4-ad49-ddbf806557a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "20af4cb7-d594-4ea4-ad49-ddbf806557a2" (UID: "20af4cb7-d594-4ea4-ad49-ddbf806557a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.004336 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20af4cb7-d594-4ea4-ad49-ddbf806557a2-kube-api-access-d2zvb" (OuterVolumeSpecName: "kube-api-access-d2zvb") pod "20af4cb7-d594-4ea4-ad49-ddbf806557a2" (UID: "20af4cb7-d594-4ea4-ad49-ddbf806557a2"). InnerVolumeSpecName "kube-api-access-d2zvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.004435 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20af4cb7-d594-4ea4-ad49-ddbf806557a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "20af4cb7-d594-4ea4-ad49-ddbf806557a2" (UID: "20af4cb7-d594-4ea4-ad49-ddbf806557a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.099549 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20af4cb7-d594-4ea4-ad49-ddbf806557a2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.099597 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2zvb\" (UniqueName: \"kubernetes.io/projected/20af4cb7-d594-4ea4-ad49-ddbf806557a2-kube-api-access-d2zvb\") on node \"crc\" DevicePath \"\"" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.099608 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20af4cb7-d594-4ea4-ad49-ddbf806557a2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.535883 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" event={"ID":"20af4cb7-d594-4ea4-ad49-ddbf806557a2","Type":"ContainerDied","Data":"766f7d481db074b2d5cb71e512b9b3f78040434b7a54acad1bd416d1b9820563"} Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.536376 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="766f7d481db074b2d5cb71e512b9b3f78040434b7a54acad1bd416d1b9820563" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.535952 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404095-dvr6c" Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.934087 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj"] Nov 27 12:15:03 crc kubenswrapper[4796]: I1127 12:15:03.941985 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404050-fzdcj"] Nov 27 12:15:05 crc kubenswrapper[4796]: I1127 12:15:05.585445 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b725a2e-30e9-455e-b359-9f0ac33eaf00" path="/var/lib/kubelet/pods/6b725a2e-30e9-455e-b359-9f0ac33eaf00/volumes" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.550488 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5sbr9"] Nov 27 12:15:25 crc kubenswrapper[4796]: E1127 12:15:25.551427 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20af4cb7-d594-4ea4-ad49-ddbf806557a2" containerName="collect-profiles" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.551443 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="20af4cb7-d594-4ea4-ad49-ddbf806557a2" containerName="collect-profiles" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.551714 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="20af4cb7-d594-4ea4-ad49-ddbf806557a2" containerName="collect-profiles" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.553237 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.625748 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sbr9"] Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.704863 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbqn5\" (UniqueName: \"kubernetes.io/projected/d6aafdf0-c037-4344-8c61-58354c2bc9ef-kube-api-access-qbqn5\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.705131 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-catalog-content\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.705342 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-utilities\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.807354 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-utilities\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.807511 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbqn5\" (UniqueName: \"kubernetes.io/projected/d6aafdf0-c037-4344-8c61-58354c2bc9ef-kube-api-access-qbqn5\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.807645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-catalog-content\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.808283 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-catalog-content\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.808601 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-utilities\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.828563 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbqn5\" (UniqueName: \"kubernetes.io/projected/d6aafdf0-c037-4344-8c61-58354c2bc9ef-kube-api-access-qbqn5\") pod \"redhat-marketplace-5sbr9\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:25 crc kubenswrapper[4796]: I1127 12:15:25.944417 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:26 crc kubenswrapper[4796]: I1127 12:15:26.306853 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sbr9"] Nov 27 12:15:26 crc kubenswrapper[4796]: I1127 12:15:26.745933 4796 generic.go:334] "Generic (PLEG): container finished" podID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerID="e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242" exitCode=0 Nov 27 12:15:26 crc kubenswrapper[4796]: I1127 12:15:26.745987 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sbr9" event={"ID":"d6aafdf0-c037-4344-8c61-58354c2bc9ef","Type":"ContainerDied","Data":"e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242"} Nov 27 12:15:26 crc kubenswrapper[4796]: I1127 12:15:26.746220 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sbr9" event={"ID":"d6aafdf0-c037-4344-8c61-58354c2bc9ef","Type":"ContainerStarted","Data":"ef67f6333c673edb908370fd031e4b583832a8800d2e4be626dca5506e907cc7"} Nov 27 12:15:27 crc kubenswrapper[4796]: I1127 12:15:27.756641 4796 generic.go:334] "Generic (PLEG): container finished" podID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerID="eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154" exitCode=0 Nov 27 12:15:27 crc kubenswrapper[4796]: I1127 12:15:27.756732 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sbr9" event={"ID":"d6aafdf0-c037-4344-8c61-58354c2bc9ef","Type":"ContainerDied","Data":"eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154"} Nov 27 12:15:28 crc kubenswrapper[4796]: I1127 12:15:28.774416 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sbr9" event={"ID":"d6aafdf0-c037-4344-8c61-58354c2bc9ef","Type":"ContainerStarted","Data":"3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7"} Nov 27 12:15:28 crc kubenswrapper[4796]: I1127 12:15:28.798731 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5sbr9" podStartSLOduration=2.28675223 podStartE2EDuration="3.798714446s" podCreationTimestamp="2025-11-27 12:15:25 +0000 UTC" firstStartedPulling="2025-11-27 12:15:26.748749848 +0000 UTC m=+3044.267068766" lastFinishedPulling="2025-11-27 12:15:28.260712074 +0000 UTC m=+3045.779030982" observedRunningTime="2025-11-27 12:15:28.79076968 +0000 UTC m=+3046.309088598" watchObservedRunningTime="2025-11-27 12:15:28.798714446 +0000 UTC m=+3046.317033364" Nov 27 12:15:35 crc kubenswrapper[4796]: I1127 12:15:35.944638 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:35 crc kubenswrapper[4796]: I1127 12:15:35.945133 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:35 crc kubenswrapper[4796]: I1127 12:15:35.993920 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:36 crc kubenswrapper[4796]: I1127 12:15:36.901946 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:36 crc kubenswrapper[4796]: I1127 12:15:36.964245 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sbr9"] Nov 27 12:15:38 crc kubenswrapper[4796]: I1127 12:15:38.861085 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5sbr9" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="registry-server" containerID="cri-o://3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7" gracePeriod=2 Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.351034 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.419615 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6aafdf0-c037-4344-8c61-58354c2bc9ef-kube-api-access-qbqn5" (OuterVolumeSpecName: "kube-api-access-qbqn5") pod "d6aafdf0-c037-4344-8c61-58354c2bc9ef" (UID: "d6aafdf0-c037-4344-8c61-58354c2bc9ef"). InnerVolumeSpecName "kube-api-access-qbqn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.426299 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbqn5\" (UniqueName: \"kubernetes.io/projected/d6aafdf0-c037-4344-8c61-58354c2bc9ef-kube-api-access-qbqn5\") pod \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.426543 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-catalog-content\") pod \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.426585 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-utilities\") pod \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\" (UID: \"d6aafdf0-c037-4344-8c61-58354c2bc9ef\") " Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.427314 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbqn5\" (UniqueName: \"kubernetes.io/projected/d6aafdf0-c037-4344-8c61-58354c2bc9ef-kube-api-access-qbqn5\") on node \"crc\" DevicePath \"\"" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.427498 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-utilities" (OuterVolumeSpecName: "utilities") pod "d6aafdf0-c037-4344-8c61-58354c2bc9ef" (UID: "d6aafdf0-c037-4344-8c61-58354c2bc9ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.445635 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6aafdf0-c037-4344-8c61-58354c2bc9ef" (UID: "d6aafdf0-c037-4344-8c61-58354c2bc9ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.530066 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.530129 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6aafdf0-c037-4344-8c61-58354c2bc9ef-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.872993 4796 generic.go:334] "Generic (PLEG): container finished" podID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerID="3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7" exitCode=0 Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.873047 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sbr9" event={"ID":"d6aafdf0-c037-4344-8c61-58354c2bc9ef","Type":"ContainerDied","Data":"3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7"} Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.873081 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sbr9" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.873108 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sbr9" event={"ID":"d6aafdf0-c037-4344-8c61-58354c2bc9ef","Type":"ContainerDied","Data":"ef67f6333c673edb908370fd031e4b583832a8800d2e4be626dca5506e907cc7"} Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.873131 4796 scope.go:117] "RemoveContainer" containerID="3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.902946 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sbr9"] Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.904643 4796 scope.go:117] "RemoveContainer" containerID="eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.910820 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sbr9"] Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.934166 4796 scope.go:117] "RemoveContainer" containerID="e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.973091 4796 scope.go:117] "RemoveContainer" containerID="3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7" Nov 27 12:15:39 crc kubenswrapper[4796]: E1127 12:15:39.973649 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7\": container with ID starting with 3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7 not found: ID does not exist" containerID="3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.973704 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7"} err="failed to get container status \"3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7\": rpc error: code = NotFound desc = could not find container \"3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7\": container with ID starting with 3e51feb64dbb9c19e7eb6c94ddf73f200d61c3d6dd0d6e1644cbe0085ed044c7 not found: ID does not exist" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.973729 4796 scope.go:117] "RemoveContainer" containerID="eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154" Nov 27 12:15:39 crc kubenswrapper[4796]: E1127 12:15:39.974188 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154\": container with ID starting with eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154 not found: ID does not exist" containerID="eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.974214 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154"} err="failed to get container status \"eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154\": rpc error: code = NotFound desc = could not find container \"eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154\": container with ID starting with eb984a964f76ae70c8a258bd24f30205225c28c17bb053bdefdfb5c7ebafc154 not found: ID does not exist" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.974228 4796 scope.go:117] "RemoveContainer" containerID="e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242" Nov 27 12:15:39 crc kubenswrapper[4796]: E1127 12:15:39.974550 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242\": container with ID starting with e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242 not found: ID does not exist" containerID="e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242" Nov 27 12:15:39 crc kubenswrapper[4796]: I1127 12:15:39.974581 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242"} err="failed to get container status \"e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242\": rpc error: code = NotFound desc = could not find container \"e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242\": container with ID starting with e440cdba5b8c36ee60af418bc080bf0bf5f5a0d80f6f8629d8686875cb863242 not found: ID does not exist" Nov 27 12:15:41 crc kubenswrapper[4796]: I1127 12:15:41.584172 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" path="/var/lib/kubelet/pods/d6aafdf0-c037-4344-8c61-58354c2bc9ef/volumes" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.432171 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b5s4c"] Nov 27 12:15:47 crc kubenswrapper[4796]: E1127 12:15:47.433234 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="extract-utilities" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.433252 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="extract-utilities" Nov 27 12:15:47 crc kubenswrapper[4796]: E1127 12:15:47.433314 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="registry-server" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.433323 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="registry-server" Nov 27 12:15:47 crc kubenswrapper[4796]: E1127 12:15:47.433346 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="extract-content" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.433354 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="extract-content" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.433606 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6aafdf0-c037-4344-8c61-58354c2bc9ef" containerName="registry-server" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.435576 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.441519 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5s4c"] Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.582963 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf5zr\" (UniqueName: \"kubernetes.io/projected/e30e01ab-fa18-4884-8dab-b40053bbcd5b-kube-api-access-jf5zr\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.583003 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-utilities\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.583162 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-catalog-content\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.684927 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-catalog-content\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.685063 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf5zr\" (UniqueName: \"kubernetes.io/projected/e30e01ab-fa18-4884-8dab-b40053bbcd5b-kube-api-access-jf5zr\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.685101 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-utilities\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.685542 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-catalog-content\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.686498 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-utilities\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.712208 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf5zr\" (UniqueName: \"kubernetes.io/projected/e30e01ab-fa18-4884-8dab-b40053bbcd5b-kube-api-access-jf5zr\") pod \"redhat-operators-b5s4c\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:47 crc kubenswrapper[4796]: I1127 12:15:47.787856 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:48 crc kubenswrapper[4796]: I1127 12:15:48.263247 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5s4c"] Nov 27 12:15:48 crc kubenswrapper[4796]: I1127 12:15:48.961937 4796 generic.go:334] "Generic (PLEG): container finished" podID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerID="e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3" exitCode=0 Nov 27 12:15:48 crc kubenswrapper[4796]: I1127 12:15:48.961974 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerDied","Data":"e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3"} Nov 27 12:15:48 crc kubenswrapper[4796]: I1127 12:15:48.962198 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerStarted","Data":"502faa0d59801d3338a7549c11c7060c00d8d6f3f5b751b1deeac081b517f57f"} Nov 27 12:15:50 crc kubenswrapper[4796]: I1127 12:15:50.983055 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerStarted","Data":"74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4"} Nov 27 12:15:52 crc kubenswrapper[4796]: I1127 12:15:52.338125 4796 scope.go:117] "RemoveContainer" containerID="7bd33a3aaa34f511ff2a890b3ed6ce4955f5f3e3c1d65214160e7fab39cbbb71" Nov 27 12:15:53 crc kubenswrapper[4796]: I1127 12:15:53.005186 4796 generic.go:334] "Generic (PLEG): container finished" podID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerID="74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4" exitCode=0 Nov 27 12:15:53 crc kubenswrapper[4796]: I1127 12:15:53.005338 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerDied","Data":"74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4"} Nov 27 12:15:55 crc kubenswrapper[4796]: I1127 12:15:55.026571 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerStarted","Data":"18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7"} Nov 27 12:15:55 crc kubenswrapper[4796]: I1127 12:15:55.052098 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b5s4c" podStartSLOduration=2.942139123 podStartE2EDuration="8.052080561s" podCreationTimestamp="2025-11-27 12:15:47 +0000 UTC" firstStartedPulling="2025-11-27 12:15:48.963894482 +0000 UTC m=+3066.482213420" lastFinishedPulling="2025-11-27 12:15:54.07383594 +0000 UTC m=+3071.592154858" observedRunningTime="2025-11-27 12:15:55.044930128 +0000 UTC m=+3072.563249046" watchObservedRunningTime="2025-11-27 12:15:55.052080561 +0000 UTC m=+3072.570399479" Nov 27 12:15:57 crc kubenswrapper[4796]: I1127 12:15:57.789748 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:57 crc kubenswrapper[4796]: I1127 12:15:57.789981 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:15:58 crc kubenswrapper[4796]: I1127 12:15:58.834154 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b5s4c" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="registry-server" probeResult="failure" output=< Nov 27 12:15:58 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:15:58 crc kubenswrapper[4796]: > Nov 27 12:16:07 crc kubenswrapper[4796]: I1127 12:16:07.858668 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:16:07 crc kubenswrapper[4796]: I1127 12:16:07.916559 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:16:08 crc kubenswrapper[4796]: I1127 12:16:08.095689 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5s4c"] Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.160953 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b5s4c" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="registry-server" containerID="cri-o://18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7" gracePeriod=2 Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.615788 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.700015 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-catalog-content\") pod \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.700081 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf5zr\" (UniqueName: \"kubernetes.io/projected/e30e01ab-fa18-4884-8dab-b40053bbcd5b-kube-api-access-jf5zr\") pod \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.700102 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-utilities\") pod \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\" (UID: \"e30e01ab-fa18-4884-8dab-b40053bbcd5b\") " Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.711466 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e30e01ab-fa18-4884-8dab-b40053bbcd5b-kube-api-access-jf5zr" (OuterVolumeSpecName: "kube-api-access-jf5zr") pod "e30e01ab-fa18-4884-8dab-b40053bbcd5b" (UID: "e30e01ab-fa18-4884-8dab-b40053bbcd5b"). InnerVolumeSpecName "kube-api-access-jf5zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.712028 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-utilities" (OuterVolumeSpecName: "utilities") pod "e30e01ab-fa18-4884-8dab-b40053bbcd5b" (UID: "e30e01ab-fa18-4884-8dab-b40053bbcd5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.793046 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e30e01ab-fa18-4884-8dab-b40053bbcd5b" (UID: "e30e01ab-fa18-4884-8dab-b40053bbcd5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.802720 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.802760 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf5zr\" (UniqueName: \"kubernetes.io/projected/e30e01ab-fa18-4884-8dab-b40053bbcd5b-kube-api-access-jf5zr\") on node \"crc\" DevicePath \"\"" Nov 27 12:16:09 crc kubenswrapper[4796]: I1127 12:16:09.802772 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30e01ab-fa18-4884-8dab-b40053bbcd5b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.172768 4796 generic.go:334] "Generic (PLEG): container finished" podID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerID="18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7" exitCode=0 Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.172831 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5s4c" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.172843 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerDied","Data":"18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7"} Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.173295 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5s4c" event={"ID":"e30e01ab-fa18-4884-8dab-b40053bbcd5b","Type":"ContainerDied","Data":"502faa0d59801d3338a7549c11c7060c00d8d6f3f5b751b1deeac081b517f57f"} Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.173319 4796 scope.go:117] "RemoveContainer" containerID="18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.199522 4796 scope.go:117] "RemoveContainer" containerID="74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.219338 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5s4c"] Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.239972 4796 scope.go:117] "RemoveContainer" containerID="e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.247607 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b5s4c"] Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.294334 4796 scope.go:117] "RemoveContainer" containerID="18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7" Nov 27 12:16:10 crc kubenswrapper[4796]: E1127 12:16:10.296029 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7\": container with ID starting with 18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7 not found: ID does not exist" containerID="18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.296069 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7"} err="failed to get container status \"18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7\": rpc error: code = NotFound desc = could not find container \"18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7\": container with ID starting with 18c8fc31d254909da72662e7ce51b96542511aeef73d879012501a65e64c35c7 not found: ID does not exist" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.296094 4796 scope.go:117] "RemoveContainer" containerID="74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4" Nov 27 12:16:10 crc kubenswrapper[4796]: E1127 12:16:10.300414 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4\": container with ID starting with 74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4 not found: ID does not exist" containerID="74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.300448 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4"} err="failed to get container status \"74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4\": rpc error: code = NotFound desc = could not find container \"74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4\": container with ID starting with 74030299e26b767b54374210e7b75115d35c770ef0a7dfcecd5322405d9429d4 not found: ID does not exist" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.300467 4796 scope.go:117] "RemoveContainer" containerID="e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3" Nov 27 12:16:10 crc kubenswrapper[4796]: E1127 12:16:10.300949 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3\": container with ID starting with e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3 not found: ID does not exist" containerID="e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3" Nov 27 12:16:10 crc kubenswrapper[4796]: I1127 12:16:10.300998 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3"} err="failed to get container status \"e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3\": rpc error: code = NotFound desc = could not find container \"e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3\": container with ID starting with e69da88df176fa5914a7153dbe0d02ac6072ebd65fdd5663cfcf071e877e92d3 not found: ID does not exist" Nov 27 12:16:11 crc kubenswrapper[4796]: I1127 12:16:11.582409 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" path="/var/lib/kubelet/pods/e30e01ab-fa18-4884-8dab-b40053bbcd5b/volumes" Nov 27 12:16:31 crc kubenswrapper[4796]: I1127 12:16:31.879317 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:16:31 crc kubenswrapper[4796]: I1127 12:16:31.879934 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:17:01 crc kubenswrapper[4796]: I1127 12:17:01.879387 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:17:01 crc kubenswrapper[4796]: I1127 12:17:01.879839 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:17:31 crc kubenswrapper[4796]: I1127 12:17:31.879838 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:17:31 crc kubenswrapper[4796]: I1127 12:17:31.880536 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:17:31 crc kubenswrapper[4796]: I1127 12:17:31.880588 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:17:31 crc kubenswrapper[4796]: I1127 12:17:31.881526 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:17:31 crc kubenswrapper[4796]: I1127 12:17:31.881599 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" gracePeriod=600 Nov 27 12:17:32 crc kubenswrapper[4796]: E1127 12:17:32.006848 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:17:32 crc kubenswrapper[4796]: I1127 12:17:32.996511 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" exitCode=0 Nov 27 12:17:32 crc kubenswrapper[4796]: I1127 12:17:32.996566 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e"} Nov 27 12:17:32 crc kubenswrapper[4796]: I1127 12:17:32.996905 4796 scope.go:117] "RemoveContainer" containerID="5836e62a2bbad31fe21baa6664ea972f4b559db15b327f1ae4158570d6c719b1" Nov 27 12:17:32 crc kubenswrapper[4796]: I1127 12:17:32.997903 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:17:32 crc kubenswrapper[4796]: E1127 12:17:32.998567 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:17:42 crc kubenswrapper[4796]: I1127 12:17:42.088715 4796 generic.go:334] "Generic (PLEG): container finished" podID="a779ff89-a04b-41a9-acfe-64e318bbf745" containerID="29a76dbd19f3d551770239a98c53d1492336de56369116ec78791b17d003112a" exitCode=0 Nov 27 12:17:42 crc kubenswrapper[4796]: I1127 12:17:42.088873 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" event={"ID":"a779ff89-a04b-41a9-acfe-64e318bbf745","Type":"ContainerDied","Data":"29a76dbd19f3d551770239a98c53d1492336de56369116ec78791b17d003112a"} Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.458560 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.587731 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-ssh-key\") pod \"a779ff89-a04b-41a9-acfe-64e318bbf745\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.588020 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-combined-ca-bundle\") pod \"a779ff89-a04b-41a9-acfe-64e318bbf745\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.588133 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjqp5\" (UniqueName: \"kubernetes.io/projected/a779ff89-a04b-41a9-acfe-64e318bbf745-kube-api-access-wjqp5\") pod \"a779ff89-a04b-41a9-acfe-64e318bbf745\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.588248 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-secret-0\") pod \"a779ff89-a04b-41a9-acfe-64e318bbf745\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.588400 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-inventory\") pod \"a779ff89-a04b-41a9-acfe-64e318bbf745\" (UID: \"a779ff89-a04b-41a9-acfe-64e318bbf745\") " Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.593802 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a779ff89-a04b-41a9-acfe-64e318bbf745-kube-api-access-wjqp5" (OuterVolumeSpecName: "kube-api-access-wjqp5") pod "a779ff89-a04b-41a9-acfe-64e318bbf745" (UID: "a779ff89-a04b-41a9-acfe-64e318bbf745"). InnerVolumeSpecName "kube-api-access-wjqp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.599505 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a779ff89-a04b-41a9-acfe-64e318bbf745" (UID: "a779ff89-a04b-41a9-acfe-64e318bbf745"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.615931 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-inventory" (OuterVolumeSpecName: "inventory") pod "a779ff89-a04b-41a9-acfe-64e318bbf745" (UID: "a779ff89-a04b-41a9-acfe-64e318bbf745"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.622198 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "a779ff89-a04b-41a9-acfe-64e318bbf745" (UID: "a779ff89-a04b-41a9-acfe-64e318bbf745"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.622709 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a779ff89-a04b-41a9-acfe-64e318bbf745" (UID: "a779ff89-a04b-41a9-acfe-64e318bbf745"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.690486 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.690536 4796 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.690549 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjqp5\" (UniqueName: \"kubernetes.io/projected/a779ff89-a04b-41a9-acfe-64e318bbf745-kube-api-access-wjqp5\") on node \"crc\" DevicePath \"\"" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.690558 4796 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:17:43 crc kubenswrapper[4796]: I1127 12:17:43.690580 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a779ff89-a04b-41a9-acfe-64e318bbf745-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.108093 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" event={"ID":"a779ff89-a04b-41a9-acfe-64e318bbf745","Type":"ContainerDied","Data":"75657f5fa5bb8a36703f9518f2719d8deb655315efb80ce97353e3d4113a9818"} Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.108137 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75657f5fa5bb8a36703f9518f2719d8deb655315efb80ce97353e3d4113a9818" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.108194 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.245012 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d"] Nov 27 12:17:44 crc kubenswrapper[4796]: E1127 12:17:44.245815 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a779ff89-a04b-41a9-acfe-64e318bbf745" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.245842 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a779ff89-a04b-41a9-acfe-64e318bbf745" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 27 12:17:44 crc kubenswrapper[4796]: E1127 12:17:44.245875 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="extract-utilities" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.245885 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="extract-utilities" Nov 27 12:17:44 crc kubenswrapper[4796]: E1127 12:17:44.245923 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="registry-server" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.245932 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="registry-server" Nov 27 12:17:44 crc kubenswrapper[4796]: E1127 12:17:44.245994 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="extract-content" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.246004 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="extract-content" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.246557 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a779ff89-a04b-41a9-acfe-64e318bbf745" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.246584 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e30e01ab-fa18-4884-8dab-b40053bbcd5b" containerName="registry-server" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.247677 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.250709 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.252941 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.253531 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.253899 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.254135 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.254545 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.254821 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.264075 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d"] Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302397 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302448 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302484 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302551 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302573 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302612 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302683 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.302800 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl6kv\" (UniqueName: \"kubernetes.io/projected/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-kube-api-access-fl6kv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405000 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405095 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405149 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405219 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl6kv\" (UniqueName: \"kubernetes.io/projected/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-kube-api-access-fl6kv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405310 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405332 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405356 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405391 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.405415 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.406192 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.409753 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.409844 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.410186 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.410643 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.410770 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.411204 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.413481 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.424050 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl6kv\" (UniqueName: \"kubernetes.io/projected/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-kube-api-access-fl6kv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-42r5d\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:44 crc kubenswrapper[4796]: I1127 12:17:44.577319 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:17:45 crc kubenswrapper[4796]: I1127 12:17:45.125351 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d"] Nov 27 12:17:45 crc kubenswrapper[4796]: I1127 12:17:45.130036 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:17:46 crc kubenswrapper[4796]: I1127 12:17:46.142416 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" event={"ID":"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941","Type":"ContainerStarted","Data":"ac4c0fd0b251caea12b6c7029c66d271dc027a0d4959562d47be706b444728fb"} Nov 27 12:17:46 crc kubenswrapper[4796]: I1127 12:17:46.142842 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" event={"ID":"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941","Type":"ContainerStarted","Data":"ff4688cc26749d8eec517aa178c6bba6ded1a66610778ede4af0cd5deeffd3e0"} Nov 27 12:17:46 crc kubenswrapper[4796]: I1127 12:17:46.179617 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" podStartSLOduration=1.720310598 podStartE2EDuration="2.179590964s" podCreationTimestamp="2025-11-27 12:17:44 +0000 UTC" firstStartedPulling="2025-11-27 12:17:45.129841903 +0000 UTC m=+3182.648160821" lastFinishedPulling="2025-11-27 12:17:45.589122269 +0000 UTC m=+3183.107441187" observedRunningTime="2025-11-27 12:17:46.171592357 +0000 UTC m=+3183.689911315" watchObservedRunningTime="2025-11-27 12:17:46.179590964 +0000 UTC m=+3183.697909892" Nov 27 12:17:47 crc kubenswrapper[4796]: I1127 12:17:47.569372 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:17:47 crc kubenswrapper[4796]: E1127 12:17:47.570045 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:17:58 crc kubenswrapper[4796]: I1127 12:17:58.570088 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:17:58 crc kubenswrapper[4796]: E1127 12:17:58.570979 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:18:12 crc kubenswrapper[4796]: I1127 12:18:12.569990 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:18:12 crc kubenswrapper[4796]: E1127 12:18:12.571444 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:18:23 crc kubenswrapper[4796]: I1127 12:18:23.575030 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:18:23 crc kubenswrapper[4796]: E1127 12:18:23.575784 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:18:35 crc kubenswrapper[4796]: I1127 12:18:35.570121 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:18:35 crc kubenswrapper[4796]: E1127 12:18:35.570908 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:18:48 crc kubenswrapper[4796]: I1127 12:18:48.569057 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:18:48 crc kubenswrapper[4796]: E1127 12:18:48.570930 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:19:00 crc kubenswrapper[4796]: I1127 12:19:00.569667 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:19:00 crc kubenswrapper[4796]: E1127 12:19:00.570672 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:19:11 crc kubenswrapper[4796]: I1127 12:19:11.569865 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:19:11 crc kubenswrapper[4796]: E1127 12:19:11.570903 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:19:24 crc kubenswrapper[4796]: I1127 12:19:24.570050 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:19:24 crc kubenswrapper[4796]: E1127 12:19:24.571547 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:19:36 crc kubenswrapper[4796]: I1127 12:19:36.570085 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:19:36 crc kubenswrapper[4796]: E1127 12:19:36.571044 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:19:50 crc kubenswrapper[4796]: I1127 12:19:50.569494 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:19:50 crc kubenswrapper[4796]: E1127 12:19:50.570337 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:20:03 crc kubenswrapper[4796]: I1127 12:20:03.575470 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:20:03 crc kubenswrapper[4796]: E1127 12:20:03.576232 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:20:14 crc kubenswrapper[4796]: I1127 12:20:14.569918 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:20:14 crc kubenswrapper[4796]: E1127 12:20:14.570603 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:20:21 crc kubenswrapper[4796]: I1127 12:20:21.532227 4796 generic.go:334] "Generic (PLEG): container finished" podID="bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" containerID="ac4c0fd0b251caea12b6c7029c66d271dc027a0d4959562d47be706b444728fb" exitCode=0 Nov 27 12:20:21 crc kubenswrapper[4796]: I1127 12:20:21.532301 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" event={"ID":"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941","Type":"ContainerDied","Data":"ac4c0fd0b251caea12b6c7029c66d271dc027a0d4959562d47be706b444728fb"} Nov 27 12:20:22 crc kubenswrapper[4796]: I1127 12:20:22.919777 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.081990 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-ssh-key\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082074 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-0\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082112 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl6kv\" (UniqueName: \"kubernetes.io/projected/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-kube-api-access-fl6kv\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082230 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-1\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082259 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-inventory\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082313 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-0\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082345 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-1\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082435 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-combined-ca-bundle\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.082486 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-extra-config-0\") pod \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\" (UID: \"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941\") " Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.089209 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.090133 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-kube-api-access-fl6kv" (OuterVolumeSpecName: "kube-api-access-fl6kv") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "kube-api-access-fl6kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.117259 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.118118 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.118586 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-inventory" (OuterVolumeSpecName: "inventory") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.119777 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.122050 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.127652 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.131544 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" (UID: "bc31c0f1-6642-4ff6-8fca-16c7ca6b4941"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185132 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185172 4796 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185189 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl6kv\" (UniqueName: \"kubernetes.io/projected/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-kube-api-access-fl6kv\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185204 4796 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185216 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185228 4796 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185240 4796 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185251 4796 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.185284 4796 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bc31c0f1-6642-4ff6-8fca-16c7ca6b4941-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.552165 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" event={"ID":"bc31c0f1-6642-4ff6-8fca-16c7ca6b4941","Type":"ContainerDied","Data":"ff4688cc26749d8eec517aa178c6bba6ded1a66610778ede4af0cd5deeffd3e0"} Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.552553 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff4688cc26749d8eec517aa178c6bba6ded1a66610778ede4af0cd5deeffd3e0" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.552521 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-42r5d" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.671666 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k"] Nov 27 12:20:23 crc kubenswrapper[4796]: E1127 12:20:23.672172 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.672197 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.672471 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc31c0f1-6642-4ff6-8fca-16c7ca6b4941" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.673158 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.677111 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.677481 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.678014 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.678062 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.679210 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dbvth" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.685216 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k"] Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.697805 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.697911 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.697935 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.697995 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.698016 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.698077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nsh2\" (UniqueName: \"kubernetes.io/projected/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-kube-api-access-9nsh2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.698161 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800289 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800393 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800437 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800488 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800510 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800556 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nsh2\" (UniqueName: \"kubernetes.io/projected/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-kube-api-access-9nsh2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.800611 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.806055 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.806081 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.806408 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.806975 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.807036 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.808193 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.823128 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nsh2\" (UniqueName: \"kubernetes.io/projected/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-kube-api-access-9nsh2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:23 crc kubenswrapper[4796]: I1127 12:20:23.992811 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:20:24 crc kubenswrapper[4796]: I1127 12:20:24.480594 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k"] Nov 27 12:20:24 crc kubenswrapper[4796]: I1127 12:20:24.566871 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" event={"ID":"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8","Type":"ContainerStarted","Data":"22e85963fcb04aa3a8fb0ffbe30106d5a52dd5b5fcee3c2c322b49c238ed84d2"} Nov 27 12:20:25 crc kubenswrapper[4796]: I1127 12:20:25.581056 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" event={"ID":"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8","Type":"ContainerStarted","Data":"40a17d240425a37e651320329dbc0521004ab546cad0b898d02c4479749dca67"} Nov 27 12:20:25 crc kubenswrapper[4796]: I1127 12:20:25.594914 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" podStartSLOduration=2.105364766 podStartE2EDuration="2.594889073s" podCreationTimestamp="2025-11-27 12:20:23 +0000 UTC" firstStartedPulling="2025-11-27 12:20:24.489899634 +0000 UTC m=+3342.008218552" lastFinishedPulling="2025-11-27 12:20:24.979423941 +0000 UTC m=+3342.497742859" observedRunningTime="2025-11-27 12:20:25.591692277 +0000 UTC m=+3343.110011205" watchObservedRunningTime="2025-11-27 12:20:25.594889073 +0000 UTC m=+3343.113207991" Nov 27 12:20:26 crc kubenswrapper[4796]: I1127 12:20:26.568485 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:20:26 crc kubenswrapper[4796]: E1127 12:20:26.568924 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:20:38 crc kubenswrapper[4796]: I1127 12:20:38.569555 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:20:38 crc kubenswrapper[4796]: E1127 12:20:38.570325 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:20:50 crc kubenswrapper[4796]: I1127 12:20:50.570939 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:20:50 crc kubenswrapper[4796]: E1127 12:20:50.571662 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:21:03 crc kubenswrapper[4796]: I1127 12:21:03.575029 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:21:03 crc kubenswrapper[4796]: E1127 12:21:03.575823 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:21:18 crc kubenswrapper[4796]: I1127 12:21:18.569196 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:21:18 crc kubenswrapper[4796]: E1127 12:21:18.570047 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:21:30 crc kubenswrapper[4796]: I1127 12:21:30.569100 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:21:30 crc kubenswrapper[4796]: E1127 12:21:30.570137 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:21:43 crc kubenswrapper[4796]: I1127 12:21:43.581427 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:21:43 crc kubenswrapper[4796]: E1127 12:21:43.582562 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.613363 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-whvtm"] Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.616596 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.632000 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whvtm"] Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.769161 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-utilities\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.769477 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-catalog-content\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.769799 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7fc9\" (UniqueName: \"kubernetes.io/projected/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-kube-api-access-l7fc9\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.871954 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7fc9\" (UniqueName: \"kubernetes.io/projected/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-kube-api-access-l7fc9\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.872434 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-utilities\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.872558 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-catalog-content\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.873013 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-utilities\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.873123 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-catalog-content\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.891969 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7fc9\" (UniqueName: \"kubernetes.io/projected/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-kube-api-access-l7fc9\") pod \"community-operators-whvtm\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:52 crc kubenswrapper[4796]: I1127 12:21:52.941547 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:21:53 crc kubenswrapper[4796]: I1127 12:21:53.488944 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whvtm"] Nov 27 12:21:54 crc kubenswrapper[4796]: I1127 12:21:54.405723 4796 generic.go:334] "Generic (PLEG): container finished" podID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerID="e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141" exitCode=0 Nov 27 12:21:54 crc kubenswrapper[4796]: I1127 12:21:54.406012 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerDied","Data":"e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141"} Nov 27 12:21:54 crc kubenswrapper[4796]: I1127 12:21:54.406050 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerStarted","Data":"5cdc645e8d21c2aa4238fc27b3d141b546652f5a1323133db01974851590b80c"} Nov 27 12:21:55 crc kubenswrapper[4796]: I1127 12:21:55.415403 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerStarted","Data":"8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde"} Nov 27 12:21:55 crc kubenswrapper[4796]: I1127 12:21:55.569119 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:21:55 crc kubenswrapper[4796]: E1127 12:21:55.569500 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:21:56 crc kubenswrapper[4796]: I1127 12:21:56.426119 4796 generic.go:334] "Generic (PLEG): container finished" podID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerID="8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde" exitCode=0 Nov 27 12:21:56 crc kubenswrapper[4796]: I1127 12:21:56.426501 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerDied","Data":"8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde"} Nov 27 12:21:57 crc kubenswrapper[4796]: I1127 12:21:57.437044 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerStarted","Data":"e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778"} Nov 27 12:21:57 crc kubenswrapper[4796]: I1127 12:21:57.459012 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-whvtm" podStartSLOduration=2.8544412770000003 podStartE2EDuration="5.458986386s" podCreationTimestamp="2025-11-27 12:21:52 +0000 UTC" firstStartedPulling="2025-11-27 12:21:54.413300794 +0000 UTC m=+3431.931619712" lastFinishedPulling="2025-11-27 12:21:57.017845903 +0000 UTC m=+3434.536164821" observedRunningTime="2025-11-27 12:21:57.454410992 +0000 UTC m=+3434.972729910" watchObservedRunningTime="2025-11-27 12:21:57.458986386 +0000 UTC m=+3434.977305314" Nov 27 12:22:02 crc kubenswrapper[4796]: I1127 12:22:02.942622 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:22:02 crc kubenswrapper[4796]: I1127 12:22:02.943012 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:22:03 crc kubenswrapper[4796]: I1127 12:22:03.003590 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:22:03 crc kubenswrapper[4796]: I1127 12:22:03.567611 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:22:05 crc kubenswrapper[4796]: I1127 12:22:05.983061 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whvtm"] Nov 27 12:22:05 crc kubenswrapper[4796]: I1127 12:22:05.983674 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-whvtm" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="registry-server" containerID="cri-o://e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778" gracePeriod=2 Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.470784 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.524627 4796 generic.go:334] "Generic (PLEG): container finished" podID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerID="e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778" exitCode=0 Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.524671 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerDied","Data":"e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778"} Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.524700 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whvtm" event={"ID":"5797fd08-3223-4f5c-bfbc-a7c0703ecfae","Type":"ContainerDied","Data":"5cdc645e8d21c2aa4238fc27b3d141b546652f5a1323133db01974851590b80c"} Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.524717 4796 scope.go:117] "RemoveContainer" containerID="e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.524729 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whvtm" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.541830 4796 scope.go:117] "RemoveContainer" containerID="8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.571167 4796 scope.go:117] "RemoveContainer" containerID="e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.606696 4796 scope.go:117] "RemoveContainer" containerID="e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778" Nov 27 12:22:06 crc kubenswrapper[4796]: E1127 12:22:06.607248 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778\": container with ID starting with e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778 not found: ID does not exist" containerID="e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.607344 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778"} err="failed to get container status \"e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778\": rpc error: code = NotFound desc = could not find container \"e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778\": container with ID starting with e00731fb1514dc10c69f76661cce75b5592f5b6755c66aa53954bad55696a778 not found: ID does not exist" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.607384 4796 scope.go:117] "RemoveContainer" containerID="8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde" Nov 27 12:22:06 crc kubenswrapper[4796]: E1127 12:22:06.607854 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde\": container with ID starting with 8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde not found: ID does not exist" containerID="8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.607897 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde"} err="failed to get container status \"8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde\": rpc error: code = NotFound desc = could not find container \"8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde\": container with ID starting with 8ecee93be4c9d2eac5f29b757204790b2a7c26c07464c01f258b14ff40d60bde not found: ID does not exist" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.607936 4796 scope.go:117] "RemoveContainer" containerID="e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141" Nov 27 12:22:06 crc kubenswrapper[4796]: E1127 12:22:06.608399 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141\": container with ID starting with e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141 not found: ID does not exist" containerID="e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.608479 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141"} err="failed to get container status \"e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141\": rpc error: code = NotFound desc = could not find container \"e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141\": container with ID starting with e64df38afa34d1a374f2c99e5766905432a58586a341fe3d9aec4888d40b0141 not found: ID does not exist" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.644178 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7fc9\" (UniqueName: \"kubernetes.io/projected/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-kube-api-access-l7fc9\") pod \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.644449 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-utilities\") pod \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.644524 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-catalog-content\") pod \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\" (UID: \"5797fd08-3223-4f5c-bfbc-a7c0703ecfae\") " Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.645637 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-utilities" (OuterVolumeSpecName: "utilities") pod "5797fd08-3223-4f5c-bfbc-a7c0703ecfae" (UID: "5797fd08-3223-4f5c-bfbc-a7c0703ecfae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.651693 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-kube-api-access-l7fc9" (OuterVolumeSpecName: "kube-api-access-l7fc9") pod "5797fd08-3223-4f5c-bfbc-a7c0703ecfae" (UID: "5797fd08-3223-4f5c-bfbc-a7c0703ecfae"). InnerVolumeSpecName "kube-api-access-l7fc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.696233 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5797fd08-3223-4f5c-bfbc-a7c0703ecfae" (UID: "5797fd08-3223-4f5c-bfbc-a7c0703ecfae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.747240 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.747323 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.747348 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7fc9\" (UniqueName: \"kubernetes.io/projected/5797fd08-3223-4f5c-bfbc-a7c0703ecfae-kube-api-access-l7fc9\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.858089 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whvtm"] Nov 27 12:22:06 crc kubenswrapper[4796]: I1127 12:22:06.873714 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-whvtm"] Nov 27 12:22:07 crc kubenswrapper[4796]: I1127 12:22:07.582532 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" path="/var/lib/kubelet/pods/5797fd08-3223-4f5c-bfbc-a7c0703ecfae/volumes" Nov 27 12:22:10 crc kubenswrapper[4796]: I1127 12:22:10.569655 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:22:10 crc kubenswrapper[4796]: E1127 12:22:10.570490 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:22:21 crc kubenswrapper[4796]: I1127 12:22:21.569288 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:22:21 crc kubenswrapper[4796]: E1127 12:22:21.570011 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:22:36 crc kubenswrapper[4796]: I1127 12:22:36.568902 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:22:37 crc kubenswrapper[4796]: I1127 12:22:37.800959 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"9b9a3460a4a4c54e56ba38ba7e96ca41227b4bbafda0b43cac4ec6cea53fcf07"} Nov 27 12:22:37 crc kubenswrapper[4796]: I1127 12:22:37.802467 4796 generic.go:334] "Generic (PLEG): container finished" podID="72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" containerID="40a17d240425a37e651320329dbc0521004ab546cad0b898d02c4479749dca67" exitCode=0 Nov 27 12:22:37 crc kubenswrapper[4796]: I1127 12:22:37.802509 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" event={"ID":"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8","Type":"ContainerDied","Data":"40a17d240425a37e651320329dbc0521004ab546cad0b898d02c4479749dca67"} Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.255346 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.393655 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-1\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.394019 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nsh2\" (UniqueName: \"kubernetes.io/projected/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-kube-api-access-9nsh2\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.394073 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-telemetry-combined-ca-bundle\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.394300 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-2\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.394400 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ssh-key\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.394448 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-inventory\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.394468 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-0\") pod \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\" (UID: \"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8\") " Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.403605 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.404101 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-kube-api-access-9nsh2" (OuterVolumeSpecName: "kube-api-access-9nsh2") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "kube-api-access-9nsh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.427645 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.431093 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.432880 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.434695 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.439962 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-inventory" (OuterVolumeSpecName: "inventory") pod "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" (UID: "72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497254 4796 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497303 4796 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497314 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497327 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497341 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nsh2\" (UniqueName: \"kubernetes.io/projected/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-kube-api-access-9nsh2\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497351 4796 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.497359 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.824213 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" event={"ID":"72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8","Type":"ContainerDied","Data":"22e85963fcb04aa3a8fb0ffbe30106d5a52dd5b5fcee3c2c322b49c238ed84d2"} Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.824252 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22e85963fcb04aa3a8fb0ffbe30106d5a52dd5b5fcee3c2c322b49c238ed84d2" Nov 27 12:22:39 crc kubenswrapper[4796]: I1127 12:22:39.824332 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.871759 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zpv7z"] Nov 27 12:22:57 crc kubenswrapper[4796]: E1127 12:22:57.872704 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="extract-utilities" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.872717 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="extract-utilities" Nov 27 12:22:57 crc kubenswrapper[4796]: E1127 12:22:57.872734 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.872741 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 27 12:22:57 crc kubenswrapper[4796]: E1127 12:22:57.872763 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="extract-content" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.872770 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="extract-content" Nov 27 12:22:57 crc kubenswrapper[4796]: E1127 12:22:57.872784 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="registry-server" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.872790 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="registry-server" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.872994 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.873010 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="5797fd08-3223-4f5c-bfbc-a7c0703ecfae" containerName="registry-server" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.874791 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:57 crc kubenswrapper[4796]: I1127 12:22:57.886387 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpv7z"] Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.048097 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/f7060db9-418e-4c2f-b122-b19a14c954f3-kube-api-access-pcxfs\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.048440 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-catalog-content\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.048500 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-utilities\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.150465 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/f7060db9-418e-4c2f-b122-b19a14c954f3-kube-api-access-pcxfs\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.150641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-catalog-content\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.150668 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-utilities\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.151177 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-catalog-content\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.151319 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-utilities\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.178252 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/f7060db9-418e-4c2f-b122-b19a14c954f3-kube-api-access-pcxfs\") pod \"certified-operators-zpv7z\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.202366 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:22:58 crc kubenswrapper[4796]: I1127 12:22:58.722649 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpv7z"] Nov 27 12:22:59 crc kubenswrapper[4796]: I1127 12:22:59.000647 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerID="80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe" exitCode=0 Nov 27 12:22:59 crc kubenswrapper[4796]: I1127 12:22:59.000943 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerDied","Data":"80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe"} Nov 27 12:22:59 crc kubenswrapper[4796]: I1127 12:22:59.000968 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerStarted","Data":"605f5bf19d46f9246a885687532d8b6b098f95b9adadde26cbd924a7e739bc6a"} Nov 27 12:22:59 crc kubenswrapper[4796]: I1127 12:22:59.003586 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:23:00 crc kubenswrapper[4796]: I1127 12:23:00.013167 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerStarted","Data":"d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a"} Nov 27 12:23:01 crc kubenswrapper[4796]: I1127 12:23:01.033804 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerID="d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a" exitCode=0 Nov 27 12:23:01 crc kubenswrapper[4796]: I1127 12:23:01.033902 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerDied","Data":"d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a"} Nov 27 12:23:03 crc kubenswrapper[4796]: I1127 12:23:03.054323 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerStarted","Data":"3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188"} Nov 27 12:23:03 crc kubenswrapper[4796]: I1127 12:23:03.075645 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zpv7z" podStartSLOduration=3.151926377 podStartE2EDuration="6.075630892s" podCreationTimestamp="2025-11-27 12:22:57 +0000 UTC" firstStartedPulling="2025-11-27 12:22:59.003380557 +0000 UTC m=+3496.521699475" lastFinishedPulling="2025-11-27 12:23:01.927085062 +0000 UTC m=+3499.445403990" observedRunningTime="2025-11-27 12:23:03.071037427 +0000 UTC m=+3500.589356345" watchObservedRunningTime="2025-11-27 12:23:03.075630892 +0000 UTC m=+3500.593949810" Nov 27 12:23:08 crc kubenswrapper[4796]: I1127 12:23:08.203362 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:23:08 crc kubenswrapper[4796]: I1127 12:23:08.204086 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:23:08 crc kubenswrapper[4796]: I1127 12:23:08.252647 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:23:09 crc kubenswrapper[4796]: I1127 12:23:09.174583 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:23:09 crc kubenswrapper[4796]: I1127 12:23:09.219946 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpv7z"] Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.142292 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zpv7z" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="registry-server" containerID="cri-o://3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188" gracePeriod=2 Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.654223 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.811636 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-catalog-content\") pod \"f7060db9-418e-4c2f-b122-b19a14c954f3\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.811829 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-utilities\") pod \"f7060db9-418e-4c2f-b122-b19a14c954f3\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.812975 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/f7060db9-418e-4c2f-b122-b19a14c954f3-kube-api-access-pcxfs\") pod \"f7060db9-418e-4c2f-b122-b19a14c954f3\" (UID: \"f7060db9-418e-4c2f-b122-b19a14c954f3\") " Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.813423 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-utilities" (OuterVolumeSpecName: "utilities") pod "f7060db9-418e-4c2f-b122-b19a14c954f3" (UID: "f7060db9-418e-4c2f-b122-b19a14c954f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.813890 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.831470 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7060db9-418e-4c2f-b122-b19a14c954f3-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "f7060db9-418e-4c2f-b122-b19a14c954f3" (UID: "f7060db9-418e-4c2f-b122-b19a14c954f3"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.870623 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7060db9-418e-4c2f-b122-b19a14c954f3" (UID: "f7060db9-418e-4c2f-b122-b19a14c954f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.915015 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7060db9-418e-4c2f-b122-b19a14c954f3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:23:11 crc kubenswrapper[4796]: I1127 12:23:11.915051 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/f7060db9-418e-4c2f-b122-b19a14c954f3-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.152245 4796 generic.go:334] "Generic (PLEG): container finished" podID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerID="3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188" exitCode=0 Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.152301 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerDied","Data":"3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188"} Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.153663 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpv7z" event={"ID":"f7060db9-418e-4c2f-b122-b19a14c954f3","Type":"ContainerDied","Data":"605f5bf19d46f9246a885687532d8b6b098f95b9adadde26cbd924a7e739bc6a"} Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.152322 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpv7z" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.153686 4796 scope.go:117] "RemoveContainer" containerID="3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.178785 4796 scope.go:117] "RemoveContainer" containerID="d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.189408 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpv7z"] Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.197687 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zpv7z"] Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.211033 4796 scope.go:117] "RemoveContainer" containerID="80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.246086 4796 scope.go:117] "RemoveContainer" containerID="3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188" Nov 27 12:23:12 crc kubenswrapper[4796]: E1127 12:23:12.246542 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188\": container with ID starting with 3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188 not found: ID does not exist" containerID="3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.246577 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188"} err="failed to get container status \"3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188\": rpc error: code = NotFound desc = could not find container \"3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188\": container with ID starting with 3da38facf405fbbcf20b04e397dffc4b56b6869e9061d43e3d9dcf3daa2de188 not found: ID does not exist" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.246600 4796 scope.go:117] "RemoveContainer" containerID="d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a" Nov 27 12:23:12 crc kubenswrapper[4796]: E1127 12:23:12.246951 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a\": container with ID starting with d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a not found: ID does not exist" containerID="d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.246979 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a"} err="failed to get container status \"d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a\": rpc error: code = NotFound desc = could not find container \"d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a\": container with ID starting with d8502daef02d3af703a79a19e46791d41bbacbd511109eef25ff31fe4f5d765a not found: ID does not exist" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.246995 4796 scope.go:117] "RemoveContainer" containerID="80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe" Nov 27 12:23:12 crc kubenswrapper[4796]: E1127 12:23:12.247190 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe\": container with ID starting with 80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe not found: ID does not exist" containerID="80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe" Nov 27 12:23:12 crc kubenswrapper[4796]: I1127 12:23:12.247222 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe"} err="failed to get container status \"80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe\": rpc error: code = NotFound desc = could not find container \"80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe\": container with ID starting with 80f0048fedfbfe81e94c66d0965f70a3a0134262d7a7d09fa4784c9bc10ea7fe not found: ID does not exist" Nov 27 12:23:13 crc kubenswrapper[4796]: I1127 12:23:13.594068 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" path="/var/lib/kubelet/pods/f7060db9-418e-4c2f-b122-b19a14c954f3/volumes" Nov 27 12:25:01 crc kubenswrapper[4796]: I1127 12:25:01.879763 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:25:01 crc kubenswrapper[4796]: I1127 12:25:01.880415 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:25:31 crc kubenswrapper[4796]: I1127 12:25:31.879382 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:25:31 crc kubenswrapper[4796]: I1127 12:25:31.880045 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:25:36 crc kubenswrapper[4796]: I1127 12:25:36.807134 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.367221 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.367903 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" containerName="openstackclient" containerID="cri-o://0cb2cdfd46668f8acf86b264ecdf74a878452ee46e557baf5cb460a79307f45f" gracePeriod=2 Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.378665 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408000 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 12:25:38 crc kubenswrapper[4796]: E1127 12:25:38.408457 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="extract-utilities" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408481 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="extract-utilities" Nov 27 12:25:38 crc kubenswrapper[4796]: E1127 12:25:38.408500 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="extract-content" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408507 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="extract-content" Nov 27 12:25:38 crc kubenswrapper[4796]: E1127 12:25:38.408529 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" containerName="openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408538 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" containerName="openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: E1127 12:25:38.408558 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="registry-server" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408565 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="registry-server" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408795 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" containerName="openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.408814 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7060db9-418e-4c2f-b122-b19a14c954f3" containerName="registry-server" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.409757 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.423014 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.441616 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.512575 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config-secret\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.512652 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.512702 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngg4g\" (UniqueName: \"kubernetes.io/projected/527ec322-c155-4c49-ad66-74c5b82c595c-kube-api-access-ngg4g\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.512881 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.613943 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config-secret\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.613987 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.614027 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngg4g\" (UniqueName: \"kubernetes.io/projected/527ec322-c155-4c49-ad66-74c5b82c595c-kube-api-access-ngg4g\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.614107 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.615130 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.620633 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.622165 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config-secret\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.634311 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngg4g\" (UniqueName: \"kubernetes.io/projected/527ec322-c155-4c49-ad66-74c5b82c595c-kube-api-access-ngg4g\") pod \"openstackclient\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " pod="openstack/openstackclient" Nov 27 12:25:38 crc kubenswrapper[4796]: I1127 12:25:38.730667 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.224567 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.447459 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-nbbj5"] Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.448973 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.461705 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-nbbj5"] Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.486603 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"527ec322-c155-4c49-ad66-74c5b82c595c","Type":"ContainerStarted","Data":"27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350"} Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.486642 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"527ec322-c155-4c49-ad66-74c5b82c595c","Type":"ContainerStarted","Data":"5de901a90b09cb51e0529135850cc5f918a6f434b13d2591c86889fdcb764a2a"} Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.508115 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.5080943869999999 podStartE2EDuration="1.508094387s" podCreationTimestamp="2025-11-27 12:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:25:39.500924323 +0000 UTC m=+3657.019243241" watchObservedRunningTime="2025-11-27 12:25:39.508094387 +0000 UTC m=+3657.026413306" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.533545 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-cbbf-account-create-update-bwk2h"] Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.534681 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.537561 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.553652 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-cbbf-account-create-update-bwk2h"] Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.636976 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g4ml\" (UniqueName: \"kubernetes.io/projected/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-kube-api-access-2g4ml\") pod \"aodh-cbbf-account-create-update-bwk2h\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.637075 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab498b8b-bcf0-4e08-82a9-ec116273269c-operator-scripts\") pod \"aodh-db-create-nbbj5\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.637099 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-operator-scripts\") pod \"aodh-cbbf-account-create-update-bwk2h\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.637174 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rszcg\" (UniqueName: \"kubernetes.io/projected/ab498b8b-bcf0-4e08-82a9-ec116273269c-kube-api-access-rszcg\") pod \"aodh-db-create-nbbj5\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.739552 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g4ml\" (UniqueName: \"kubernetes.io/projected/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-kube-api-access-2g4ml\") pod \"aodh-cbbf-account-create-update-bwk2h\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.739632 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-operator-scripts\") pod \"aodh-cbbf-account-create-update-bwk2h\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.739668 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab498b8b-bcf0-4e08-82a9-ec116273269c-operator-scripts\") pod \"aodh-db-create-nbbj5\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.739716 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rszcg\" (UniqueName: \"kubernetes.io/projected/ab498b8b-bcf0-4e08-82a9-ec116273269c-kube-api-access-rszcg\") pod \"aodh-db-create-nbbj5\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.740542 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab498b8b-bcf0-4e08-82a9-ec116273269c-operator-scripts\") pod \"aodh-db-create-nbbj5\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.740613 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-operator-scripts\") pod \"aodh-cbbf-account-create-update-bwk2h\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.756025 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rszcg\" (UniqueName: \"kubernetes.io/projected/ab498b8b-bcf0-4e08-82a9-ec116273269c-kube-api-access-rszcg\") pod \"aodh-db-create-nbbj5\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.757722 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g4ml\" (UniqueName: \"kubernetes.io/projected/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-kube-api-access-2g4ml\") pod \"aodh-cbbf-account-create-update-bwk2h\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.781023 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:39 crc kubenswrapper[4796]: I1127 12:25:39.859546 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.228043 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-nbbj5"] Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.341207 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-cbbf-account-create-update-bwk2h"] Nov 27 12:25:40 crc kubenswrapper[4796]: W1127 12:25:40.346257 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafe8e4fe_1791_461c_a9bc_bb9bd352d7d9.slice/crio-afa66f29a21c62f2c3a451df4f3c919ebc35256735433ad99a7d5c1ed8310211 WatchSource:0}: Error finding container afa66f29a21c62f2c3a451df4f3c919ebc35256735433ad99a7d5c1ed8310211: Status 404 returned error can't find the container with id afa66f29a21c62f2c3a451df4f3c919ebc35256735433ad99a7d5c1ed8310211 Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.499905 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nbbj5" event={"ID":"ab498b8b-bcf0-4e08-82a9-ec116273269c","Type":"ContainerStarted","Data":"263ba999cfc7e9c5722643ea26eeef99004d6d67e580a46ddcab4bc20c2b482d"} Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.499965 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nbbj5" event={"ID":"ab498b8b-bcf0-4e08-82a9-ec116273269c","Type":"ContainerStarted","Data":"d7dd6cfbb18fdd44fc9ab856950404865c88eabd2ae36b1d1205b663077d5f3e"} Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.505913 4796 generic.go:334] "Generic (PLEG): container finished" podID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" containerID="0cb2cdfd46668f8acf86b264ecdf74a878452ee46e557baf5cb460a79307f45f" exitCode=137 Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.508584 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cbbf-account-create-update-bwk2h" event={"ID":"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9","Type":"ContainerStarted","Data":"afa66f29a21c62f2c3a451df4f3c919ebc35256735433ad99a7d5c1ed8310211"} Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.518704 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-nbbj5" podStartSLOduration=1.5186799500000001 podStartE2EDuration="1.51867995s" podCreationTimestamp="2025-11-27 12:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:25:40.513721515 +0000 UTC m=+3658.032040433" watchObservedRunningTime="2025-11-27 12:25:40.51867995 +0000 UTC m=+3658.036998868" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.620060 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.757801 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-combined-ca-bundle\") pod \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.758320 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config-secret\") pod \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.758514 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zclr5\" (UniqueName: \"kubernetes.io/projected/c54eb17e-bce8-4e91-b9f2-766040bcfc89-kube-api-access-zclr5\") pod \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.758551 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config\") pod \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\" (UID: \"c54eb17e-bce8-4e91-b9f2-766040bcfc89\") " Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.763240 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54eb17e-bce8-4e91-b9f2-766040bcfc89-kube-api-access-zclr5" (OuterVolumeSpecName: "kube-api-access-zclr5") pod "c54eb17e-bce8-4e91-b9f2-766040bcfc89" (UID: "c54eb17e-bce8-4e91-b9f2-766040bcfc89"). InnerVolumeSpecName "kube-api-access-zclr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.788007 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c54eb17e-bce8-4e91-b9f2-766040bcfc89" (UID: "c54eb17e-bce8-4e91-b9f2-766040bcfc89"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.839349 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c54eb17e-bce8-4e91-b9f2-766040bcfc89" (UID: "c54eb17e-bce8-4e91-b9f2-766040bcfc89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.864658 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zclr5\" (UniqueName: \"kubernetes.io/projected/c54eb17e-bce8-4e91-b9f2-766040bcfc89-kube-api-access-zclr5\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.864698 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.864707 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.890449 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c54eb17e-bce8-4e91-b9f2-766040bcfc89" (UID: "c54eb17e-bce8-4e91-b9f2-766040bcfc89"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:25:40 crc kubenswrapper[4796]: I1127 12:25:40.966627 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c54eb17e-bce8-4e91-b9f2-766040bcfc89-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.517992 4796 generic.go:334] "Generic (PLEG): container finished" podID="afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" containerID="59019cb781f6b79431e6ad89679a8a559b4d315656201884fda1e550b20a5d14" exitCode=0 Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.518295 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cbbf-account-create-update-bwk2h" event={"ID":"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9","Type":"ContainerDied","Data":"59019cb781f6b79431e6ad89679a8a559b4d315656201884fda1e550b20a5d14"} Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.520195 4796 generic.go:334] "Generic (PLEG): container finished" podID="ab498b8b-bcf0-4e08-82a9-ec116273269c" containerID="263ba999cfc7e9c5722643ea26eeef99004d6d67e580a46ddcab4bc20c2b482d" exitCode=0 Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.520367 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nbbj5" event={"ID":"ab498b8b-bcf0-4e08-82a9-ec116273269c","Type":"ContainerDied","Data":"263ba999cfc7e9c5722643ea26eeef99004d6d67e580a46ddcab4bc20c2b482d"} Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.523720 4796 scope.go:117] "RemoveContainer" containerID="0cb2cdfd46668f8acf86b264ecdf74a878452ee46e557baf5cb460a79307f45f" Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.523764 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.552278 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" Nov 27 12:25:41 crc kubenswrapper[4796]: I1127 12:25:41.578994 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c54eb17e-bce8-4e91-b9f2-766040bcfc89" path="/var/lib/kubelet/pods/c54eb17e-bce8-4e91-b9f2-766040bcfc89/volumes" Nov 27 12:25:42 crc kubenswrapper[4796]: I1127 12:25:42.906699 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:42 crc kubenswrapper[4796]: I1127 12:25:42.915193 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.108517 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab498b8b-bcf0-4e08-82a9-ec116273269c-operator-scripts\") pod \"ab498b8b-bcf0-4e08-82a9-ec116273269c\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.108635 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rszcg\" (UniqueName: \"kubernetes.io/projected/ab498b8b-bcf0-4e08-82a9-ec116273269c-kube-api-access-rszcg\") pod \"ab498b8b-bcf0-4e08-82a9-ec116273269c\" (UID: \"ab498b8b-bcf0-4e08-82a9-ec116273269c\") " Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.108697 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g4ml\" (UniqueName: \"kubernetes.io/projected/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-kube-api-access-2g4ml\") pod \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.108773 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-operator-scripts\") pod \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\" (UID: \"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9\") " Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.109014 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab498b8b-bcf0-4e08-82a9-ec116273269c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab498b8b-bcf0-4e08-82a9-ec116273269c" (UID: "ab498b8b-bcf0-4e08-82a9-ec116273269c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.109717 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" (UID: "afe8e4fe-1791-461c-a9bc-bb9bd352d7d9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.109899 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab498b8b-bcf0-4e08-82a9-ec116273269c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.109938 4796 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.115939 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-kube-api-access-2g4ml" (OuterVolumeSpecName: "kube-api-access-2g4ml") pod "afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" (UID: "afe8e4fe-1791-461c-a9bc-bb9bd352d7d9"). InnerVolumeSpecName "kube-api-access-2g4ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.116075 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab498b8b-bcf0-4e08-82a9-ec116273269c-kube-api-access-rszcg" (OuterVolumeSpecName: "kube-api-access-rszcg") pod "ab498b8b-bcf0-4e08-82a9-ec116273269c" (UID: "ab498b8b-bcf0-4e08-82a9-ec116273269c"). InnerVolumeSpecName "kube-api-access-rszcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.214061 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rszcg\" (UniqueName: \"kubernetes.io/projected/ab498b8b-bcf0-4e08-82a9-ec116273269c-kube-api-access-rszcg\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.214113 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g4ml\" (UniqueName: \"kubernetes.io/projected/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9-kube-api-access-2g4ml\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.544234 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cbbf-account-create-update-bwk2h" event={"ID":"afe8e4fe-1791-461c-a9bc-bb9bd352d7d9","Type":"ContainerDied","Data":"afa66f29a21c62f2c3a451df4f3c919ebc35256735433ad99a7d5c1ed8310211"} Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.544277 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cbbf-account-create-update-bwk2h" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.544284 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afa66f29a21c62f2c3a451df4f3c919ebc35256735433ad99a7d5c1ed8310211" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.546737 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-nbbj5" Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.549009 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-nbbj5" event={"ID":"ab498b8b-bcf0-4e08-82a9-ec116273269c","Type":"ContainerDied","Data":"d7dd6cfbb18fdd44fc9ab856950404865c88eabd2ae36b1d1205b663077d5f3e"} Nov 27 12:25:43 crc kubenswrapper[4796]: I1127 12:25:43.549085 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7dd6cfbb18fdd44fc9ab856950404865c88eabd2ae36b1d1205b663077d5f3e" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.988624 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-rhjzb"] Nov 27 12:25:44 crc kubenswrapper[4796]: E1127 12:25:44.989676 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab498b8b-bcf0-4e08-82a9-ec116273269c" containerName="mariadb-database-create" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.989698 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab498b8b-bcf0-4e08-82a9-ec116273269c" containerName="mariadb-database-create" Nov 27 12:25:44 crc kubenswrapper[4796]: E1127 12:25:44.989744 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" containerName="mariadb-account-create-update" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.989753 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" containerName="mariadb-account-create-update" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.990000 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab498b8b-bcf0-4e08-82a9-ec116273269c" containerName="mariadb-database-create" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.990030 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" containerName="mariadb-account-create-update" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.990893 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.994193 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.994192 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vs2ld" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.999017 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 27 12:25:44 crc kubenswrapper[4796]: I1127 12:25:44.999627 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-rhjzb"] Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.002071 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.045063 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-combined-ca-bundle\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.045223 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw84k\" (UniqueName: \"kubernetes.io/projected/bb7aba55-765f-4264-b1c7-803655588e9a-kube-api-access-kw84k\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.045355 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-config-data\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.045409 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-scripts\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.146933 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-config-data\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.147053 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-scripts\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.147317 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-combined-ca-bundle\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.147469 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw84k\" (UniqueName: \"kubernetes.io/projected/bb7aba55-765f-4264-b1c7-803655588e9a-kube-api-access-kw84k\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.152698 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-config-data\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.153102 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-combined-ca-bundle\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.154048 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-scripts\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.173593 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw84k\" (UniqueName: \"kubernetes.io/projected/bb7aba55-765f-4264-b1c7-803655588e9a-kube-api-access-kw84k\") pod \"aodh-db-sync-rhjzb\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.315779 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:45 crc kubenswrapper[4796]: I1127 12:25:45.755211 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-rhjzb"] Nov 27 12:25:46 crc kubenswrapper[4796]: I1127 12:25:46.577312 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rhjzb" event={"ID":"bb7aba55-765f-4264-b1c7-803655588e9a","Type":"ContainerStarted","Data":"5f6191f010731db4bec4a13196a6d69f2d5a5e25591d6442adae44263621b2f9"} Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.582589 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m2mw5"] Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.585943 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.606583 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m2mw5"] Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.713807 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-utilities\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.713878 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82skh\" (UniqueName: \"kubernetes.io/projected/169ec2e6-4704-4425-9710-c5b4af3a609c-kube-api-access-82skh\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.713941 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-catalog-content\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.815898 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-utilities\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.815959 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82skh\" (UniqueName: \"kubernetes.io/projected/169ec2e6-4704-4425-9710-c5b4af3a609c-kube-api-access-82skh\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.816002 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-catalog-content\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.816611 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-catalog-content\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.816669 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-utilities\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.838589 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82skh\" (UniqueName: \"kubernetes.io/projected/169ec2e6-4704-4425-9710-c5b4af3a609c-kube-api-access-82skh\") pod \"redhat-operators-m2mw5\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:48 crc kubenswrapper[4796]: I1127 12:25:48.919380 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:50 crc kubenswrapper[4796]: W1127 12:25:50.011900 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod169ec2e6_4704_4425_9710_c5b4af3a609c.slice/crio-b4e01eb29cee89872cc93acb0dc8c12bd7cff073b8fc6f59221a697815d004a1 WatchSource:0}: Error finding container b4e01eb29cee89872cc93acb0dc8c12bd7cff073b8fc6f59221a697815d004a1: Status 404 returned error can't find the container with id b4e01eb29cee89872cc93acb0dc8c12bd7cff073b8fc6f59221a697815d004a1 Nov 27 12:25:50 crc kubenswrapper[4796]: I1127 12:25:50.014044 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m2mw5"] Nov 27 12:25:50 crc kubenswrapper[4796]: I1127 12:25:50.621231 4796 generic.go:334] "Generic (PLEG): container finished" podID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerID="005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549" exitCode=0 Nov 27 12:25:50 crc kubenswrapper[4796]: I1127 12:25:50.621923 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m2mw5" event={"ID":"169ec2e6-4704-4425-9710-c5b4af3a609c","Type":"ContainerDied","Data":"005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549"} Nov 27 12:25:50 crc kubenswrapper[4796]: I1127 12:25:50.622518 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m2mw5" event={"ID":"169ec2e6-4704-4425-9710-c5b4af3a609c","Type":"ContainerStarted","Data":"b4e01eb29cee89872cc93acb0dc8c12bd7cff073b8fc6f59221a697815d004a1"} Nov 27 12:25:50 crc kubenswrapper[4796]: I1127 12:25:50.627134 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rhjzb" event={"ID":"bb7aba55-765f-4264-b1c7-803655588e9a","Type":"ContainerStarted","Data":"9ed5b99cec6c084e247a7bba7eccfd000c31e8c094e4b7282a667c934712bce2"} Nov 27 12:25:50 crc kubenswrapper[4796]: I1127 12:25:50.672036 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-rhjzb" podStartSLOduration=2.750094463 podStartE2EDuration="6.672019908s" podCreationTimestamp="2025-11-27 12:25:44 +0000 UTC" firstStartedPulling="2025-11-27 12:25:45.759433203 +0000 UTC m=+3663.277752121" lastFinishedPulling="2025-11-27 12:25:49.681358648 +0000 UTC m=+3667.199677566" observedRunningTime="2025-11-27 12:25:50.66548472 +0000 UTC m=+3668.183803638" watchObservedRunningTime="2025-11-27 12:25:50.672019908 +0000 UTC m=+3668.190338826" Nov 27 12:25:52 crc kubenswrapper[4796]: I1127 12:25:52.649834 4796 generic.go:334] "Generic (PLEG): container finished" podID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerID="74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8" exitCode=0 Nov 27 12:25:52 crc kubenswrapper[4796]: I1127 12:25:52.649927 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m2mw5" event={"ID":"169ec2e6-4704-4425-9710-c5b4af3a609c","Type":"ContainerDied","Data":"74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8"} Nov 27 12:25:52 crc kubenswrapper[4796]: I1127 12:25:52.652617 4796 generic.go:334] "Generic (PLEG): container finished" podID="bb7aba55-765f-4264-b1c7-803655588e9a" containerID="9ed5b99cec6c084e247a7bba7eccfd000c31e8c094e4b7282a667c934712bce2" exitCode=0 Nov 27 12:25:52 crc kubenswrapper[4796]: I1127 12:25:52.652644 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rhjzb" event={"ID":"bb7aba55-765f-4264-b1c7-803655588e9a","Type":"ContainerDied","Data":"9ed5b99cec6c084e247a7bba7eccfd000c31e8c094e4b7282a667c934712bce2"} Nov 27 12:25:53 crc kubenswrapper[4796]: I1127 12:25:53.669079 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m2mw5" event={"ID":"169ec2e6-4704-4425-9710-c5b4af3a609c","Type":"ContainerStarted","Data":"19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2"} Nov 27 12:25:53 crc kubenswrapper[4796]: I1127 12:25:53.695878 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m2mw5" podStartSLOduration=3.181384955 podStartE2EDuration="5.695859801s" podCreationTimestamp="2025-11-27 12:25:48 +0000 UTC" firstStartedPulling="2025-11-27 12:25:50.623488598 +0000 UTC m=+3668.141807646" lastFinishedPulling="2025-11-27 12:25:53.137963544 +0000 UTC m=+3670.656282492" observedRunningTime="2025-11-27 12:25:53.692789338 +0000 UTC m=+3671.211108276" watchObservedRunningTime="2025-11-27 12:25:53.695859801 +0000 UTC m=+3671.214178719" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.677049 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-rhjzb" event={"ID":"bb7aba55-765f-4264-b1c7-803655588e9a","Type":"ContainerDied","Data":"5f6191f010731db4bec4a13196a6d69f2d5a5e25591d6442adae44263621b2f9"} Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.677471 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f6191f010731db4bec4a13196a6d69f2d5a5e25591d6442adae44263621b2f9" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.734202 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.832620 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-config-data\") pod \"bb7aba55-765f-4264-b1c7-803655588e9a\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.832676 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-scripts\") pod \"bb7aba55-765f-4264-b1c7-803655588e9a\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.832733 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-combined-ca-bundle\") pod \"bb7aba55-765f-4264-b1c7-803655588e9a\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.832882 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw84k\" (UniqueName: \"kubernetes.io/projected/bb7aba55-765f-4264-b1c7-803655588e9a-kube-api-access-kw84k\") pod \"bb7aba55-765f-4264-b1c7-803655588e9a\" (UID: \"bb7aba55-765f-4264-b1c7-803655588e9a\") " Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.837735 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-scripts" (OuterVolumeSpecName: "scripts") pod "bb7aba55-765f-4264-b1c7-803655588e9a" (UID: "bb7aba55-765f-4264-b1c7-803655588e9a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.839180 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7aba55-765f-4264-b1c7-803655588e9a-kube-api-access-kw84k" (OuterVolumeSpecName: "kube-api-access-kw84k") pod "bb7aba55-765f-4264-b1c7-803655588e9a" (UID: "bb7aba55-765f-4264-b1c7-803655588e9a"). InnerVolumeSpecName "kube-api-access-kw84k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.865053 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb7aba55-765f-4264-b1c7-803655588e9a" (UID: "bb7aba55-765f-4264-b1c7-803655588e9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.867068 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-config-data" (OuterVolumeSpecName: "config-data") pod "bb7aba55-765f-4264-b1c7-803655588e9a" (UID: "bb7aba55-765f-4264-b1c7-803655588e9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.935379 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.935417 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw84k\" (UniqueName: \"kubernetes.io/projected/bb7aba55-765f-4264-b1c7-803655588e9a-kube-api-access-kw84k\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.935428 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:54 crc kubenswrapper[4796]: I1127 12:25:54.935436 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb7aba55-765f-4264-b1c7-803655588e9a-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:25:55 crc kubenswrapper[4796]: I1127 12:25:55.684780 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-rhjzb" Nov 27 12:25:58 crc kubenswrapper[4796]: I1127 12:25:58.919963 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:58 crc kubenswrapper[4796]: I1127 12:25:58.920342 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:25:59 crc kubenswrapper[4796]: I1127 12:25:59.963325 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m2mw5" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="registry-server" probeResult="failure" output=< Nov 27 12:25:59 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:25:59 crc kubenswrapper[4796]: > Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.098781 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:00 crc kubenswrapper[4796]: E1127 12:26:00.099481 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7aba55-765f-4264-b1c7-803655588e9a" containerName="aodh-db-sync" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.099503 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7aba55-765f-4264-b1c7-803655588e9a" containerName="aodh-db-sync" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.099706 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7aba55-765f-4264-b1c7-803655588e9a" containerName="aodh-db-sync" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.101356 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.103762 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.103927 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vs2ld" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.107620 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.142078 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.248949 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.249008 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-config-data\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.249188 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-scripts\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.249481 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52zxd\" (UniqueName: \"kubernetes.io/projected/547e11ed-da55-4baa-b1f1-737907bd7ea1-kube-api-access-52zxd\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.351920 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-scripts\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.352038 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52zxd\" (UniqueName: \"kubernetes.io/projected/547e11ed-da55-4baa-b1f1-737907bd7ea1-kube-api-access-52zxd\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.352099 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.352136 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-config-data\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.358525 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-scripts\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.360916 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.370868 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52zxd\" (UniqueName: \"kubernetes.io/projected/547e11ed-da55-4baa-b1f1-737907bd7ea1-kube-api-access-52zxd\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.371215 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-config-data\") pod \"aodh-0\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " pod="openstack/aodh-0" Nov 27 12:26:00 crc kubenswrapper[4796]: I1127 12:26:00.436756 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:26:01 crc kubenswrapper[4796]: I1127 12:26:01.879609 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:26:01 crc kubenswrapper[4796]: I1127 12:26:01.880398 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:26:01 crc kubenswrapper[4796]: I1127 12:26:01.880748 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:26:01 crc kubenswrapper[4796]: I1127 12:26:01.881724 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b9a3460a4a4c54e56ba38ba7e96ca41227b4bbafda0b43cac4ec6cea53fcf07"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:26:01 crc kubenswrapper[4796]: I1127 12:26:01.881817 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://9b9a3460a4a4c54e56ba38ba7e96ca41227b4bbafda0b43cac4ec6cea53fcf07" gracePeriod=600 Nov 27 12:26:01 crc kubenswrapper[4796]: I1127 12:26:01.954671 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:02 crc kubenswrapper[4796]: I1127 12:26:02.751660 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerStarted","Data":"3067a3c162f82f285d1c3d794a09aacd83c7bdc60dbc6289ecef04378be99bbc"} Nov 27 12:26:02 crc kubenswrapper[4796]: I1127 12:26:02.755197 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="9b9a3460a4a4c54e56ba38ba7e96ca41227b4bbafda0b43cac4ec6cea53fcf07" exitCode=0 Nov 27 12:26:02 crc kubenswrapper[4796]: I1127 12:26:02.755249 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"9b9a3460a4a4c54e56ba38ba7e96ca41227b4bbafda0b43cac4ec6cea53fcf07"} Nov 27 12:26:02 crc kubenswrapper[4796]: I1127 12:26:02.755330 4796 scope.go:117] "RemoveContainer" containerID="a63d155d470aefff93d6c139e5bc43ef842fbdb3a82dd23f705d4e584ffb6a5e" Nov 27 12:26:03 crc kubenswrapper[4796]: I1127 12:26:03.636692 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:03 crc kubenswrapper[4796]: I1127 12:26:03.766298 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58"} Nov 27 12:26:04 crc kubenswrapper[4796]: I1127 12:26:04.775769 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerStarted","Data":"cf65a6d196417dd9e0d215320bb1ccbf54d8bf4a840fcfd1b73d13e55c5fff75"} Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.184826 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.186782 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-central-agent" containerID="cri-o://9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8" gracePeriod=30 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.188181 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="sg-core" containerID="cri-o://4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132" gracePeriod=30 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.188364 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="proxy-httpd" containerID="cri-o://66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664" gracePeriod=30 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.188506 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-notification-agent" containerID="cri-o://14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d" gracePeriod=30 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.786941 4796 generic.go:334] "Generic (PLEG): container finished" podID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerID="66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664" exitCode=0 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.787425 4796 generic.go:334] "Generic (PLEG): container finished" podID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerID="4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132" exitCode=2 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.787437 4796 generic.go:334] "Generic (PLEG): container finished" podID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerID="9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8" exitCode=0 Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.787202 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerDied","Data":"66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664"} Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.787470 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerDied","Data":"4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132"} Nov 27 12:26:05 crc kubenswrapper[4796]: I1127 12:26:05.787485 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerDied","Data":"9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8"} Nov 27 12:26:06 crc kubenswrapper[4796]: I1127 12:26:06.797483 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerStarted","Data":"e12be2d2ee731c35bc837e77c016f7dc5bf36321c575b9261b1af298c2d16bac"} Nov 27 12:26:07 crc kubenswrapper[4796]: I1127 12:26:07.810389 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerStarted","Data":"09fe631702ec34e296b71b34910ee8356f8e025f59efbf35d00700d7b930a870"} Nov 27 12:26:08 crc kubenswrapper[4796]: I1127 12:26:08.964603 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.017231 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.200476 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m2mw5"] Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.829909 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerStarted","Data":"18d27670bca6efcc8b66e4a40592b92bb831f2318ec4513ea5dc3dd828188684"} Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.830069 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-notifier" containerID="cri-o://09fe631702ec34e296b71b34910ee8356f8e025f59efbf35d00700d7b930a870" gracePeriod=30 Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.830073 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-listener" containerID="cri-o://18d27670bca6efcc8b66e4a40592b92bb831f2318ec4513ea5dc3dd828188684" gracePeriod=30 Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.830092 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-evaluator" containerID="cri-o://e12be2d2ee731c35bc837e77c016f7dc5bf36321c575b9261b1af298c2d16bac" gracePeriod=30 Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.830027 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-api" containerID="cri-o://cf65a6d196417dd9e0d215320bb1ccbf54d8bf4a840fcfd1b73d13e55c5fff75" gracePeriod=30 Nov 27 12:26:09 crc kubenswrapper[4796]: I1127 12:26:09.855809 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.490581158 podStartE2EDuration="9.855787826s" podCreationTimestamp="2025-11-27 12:26:00 +0000 UTC" firstStartedPulling="2025-11-27 12:26:01.957237999 +0000 UTC m=+3679.475556917" lastFinishedPulling="2025-11-27 12:26:09.322444667 +0000 UTC m=+3686.840763585" observedRunningTime="2025-11-27 12:26:09.851223222 +0000 UTC m=+3687.369542140" watchObservedRunningTime="2025-11-27 12:26:09.855787826 +0000 UTC m=+3687.374106744" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.482144 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.669778 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-scripts\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670123 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-sg-core-conf-yaml\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670169 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-log-httpd\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670224 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-config-data\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670288 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2crcj\" (UniqueName: \"kubernetes.io/projected/afc78f25-c3cb-4627-be10-a6838c84d61f-kube-api-access-2crcj\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670499 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-combined-ca-bundle\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670627 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-ceilometer-tls-certs\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.670682 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-run-httpd\") pod \"afc78f25-c3cb-4627-be10-a6838c84d61f\" (UID: \"afc78f25-c3cb-4627-be10-a6838c84d61f\") " Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.673538 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.673883 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.678455 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-scripts" (OuterVolumeSpecName: "scripts") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.682508 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc78f25-c3cb-4627-be10-a6838c84d61f-kube-api-access-2crcj" (OuterVolumeSpecName: "kube-api-access-2crcj") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "kube-api-access-2crcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.709660 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.731652 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.751694 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772875 4796 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772906 4796 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772918 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2crcj\" (UniqueName: \"kubernetes.io/projected/afc78f25-c3cb-4627-be10-a6838c84d61f-kube-api-access-2crcj\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772932 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772944 4796 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772956 4796 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/afc78f25-c3cb-4627-be10-a6838c84d61f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.772966 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.789701 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-config-data" (OuterVolumeSpecName: "config-data") pod "afc78f25-c3cb-4627-be10-a6838c84d61f" (UID: "afc78f25-c3cb-4627-be10-a6838c84d61f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.840368 4796 generic.go:334] "Generic (PLEG): container finished" podID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerID="09fe631702ec34e296b71b34910ee8356f8e025f59efbf35d00700d7b930a870" exitCode=0 Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.840423 4796 generic.go:334] "Generic (PLEG): container finished" podID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerID="e12be2d2ee731c35bc837e77c016f7dc5bf36321c575b9261b1af298c2d16bac" exitCode=0 Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.840432 4796 generic.go:334] "Generic (PLEG): container finished" podID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerID="cf65a6d196417dd9e0d215320bb1ccbf54d8bf4a840fcfd1b73d13e55c5fff75" exitCode=0 Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.840477 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerDied","Data":"09fe631702ec34e296b71b34910ee8356f8e025f59efbf35d00700d7b930a870"} Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.840507 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerDied","Data":"e12be2d2ee731c35bc837e77c016f7dc5bf36321c575b9261b1af298c2d16bac"} Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.840517 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerDied","Data":"cf65a6d196417dd9e0d215320bb1ccbf54d8bf4a840fcfd1b73d13e55c5fff75"} Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.845103 4796 generic.go:334] "Generic (PLEG): container finished" podID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerID="14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d" exitCode=0 Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.845391 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m2mw5" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="registry-server" containerID="cri-o://19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2" gracePeriod=2 Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.845824 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.845892 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerDied","Data":"14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d"} Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.847873 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"afc78f25-c3cb-4627-be10-a6838c84d61f","Type":"ContainerDied","Data":"55c2f0680d2d54bb4f7fb83c6c998b02b8141f5e0f2f28fab4d35662bb70c576"} Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.847986 4796 scope.go:117] "RemoveContainer" containerID="66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.870493 4796 scope.go:117] "RemoveContainer" containerID="4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.875074 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc78f25-c3cb-4627-be10-a6838c84d61f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.904219 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.906876 4796 scope.go:117] "RemoveContainer" containerID="14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.920911 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.934740 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.935289 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-notification-agent" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935305 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-notification-agent" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.935341 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-central-agent" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935350 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-central-agent" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.935367 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="proxy-httpd" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935376 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="proxy-httpd" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.935389 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="sg-core" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935397 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="sg-core" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935651 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="sg-core" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935678 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="proxy-httpd" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935695 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-central-agent" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.935709 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" containerName="ceilometer-notification-agent" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.937887 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.946170 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.959478 4796 scope.go:117] "RemoveContainer" containerID="9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.960922 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.961224 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.961994 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.990075 4796 scope.go:117] "RemoveContainer" containerID="66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.991561 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664\": container with ID starting with 66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664 not found: ID does not exist" containerID="66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.991594 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664"} err="failed to get container status \"66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664\": rpc error: code = NotFound desc = could not find container \"66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664\": container with ID starting with 66d52cb436b3422e1b14bd4d09504f5a6015da2c1159ecf17b7ef2a188e94664 not found: ID does not exist" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.991621 4796 scope.go:117] "RemoveContainer" containerID="4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.991928 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132\": container with ID starting with 4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132 not found: ID does not exist" containerID="4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.991956 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132"} err="failed to get container status \"4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132\": rpc error: code = NotFound desc = could not find container \"4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132\": container with ID starting with 4b93b01b25a4b4a0d9e00b5a05fc5eb4d6bc00661d25ee0fed7818360d0a5132 not found: ID does not exist" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.991978 4796 scope.go:117] "RemoveContainer" containerID="14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.992490 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d\": container with ID starting with 14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d not found: ID does not exist" containerID="14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.992556 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d"} err="failed to get container status \"14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d\": rpc error: code = NotFound desc = could not find container \"14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d\": container with ID starting with 14e3af7ebdaa2447fa624ff313672aaefc7927e6c7f9320feaa36e4e5a95027d not found: ID does not exist" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.992589 4796 scope.go:117] "RemoveContainer" containerID="9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8" Nov 27 12:26:10 crc kubenswrapper[4796]: E1127 12:26:10.992896 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8\": container with ID starting with 9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8 not found: ID does not exist" containerID="9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8" Nov 27 12:26:10 crc kubenswrapper[4796]: I1127 12:26:10.992929 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8"} err="failed to get container status \"9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8\": rpc error: code = NotFound desc = could not find container \"9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8\": container with ID starting with 9f72a6c789b8c4b3752e58a06c1693c29ad872bc2d125cb578473a4582535ba8 not found: ID does not exist" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.081634 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eeb4962a-1819-4562-baf3-1b27c476d5f1-run-httpd\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.081703 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-scripts\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.081841 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eeb4962a-1819-4562-baf3-1b27c476d5f1-log-httpd\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.081913 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-config-data\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.082365 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.082420 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.082445 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d8vc\" (UniqueName: \"kubernetes.io/projected/eeb4962a-1819-4562-baf3-1b27c476d5f1-kube-api-access-7d8vc\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.082552 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.184365 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-scripts\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.184699 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eeb4962a-1819-4562-baf3-1b27c476d5f1-log-httpd\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.184745 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-config-data\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.184948 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.184997 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d8vc\" (UniqueName: \"kubernetes.io/projected/eeb4962a-1819-4562-baf3-1b27c476d5f1-kube-api-access-7d8vc\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.185020 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.185091 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.185170 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eeb4962a-1819-4562-baf3-1b27c476d5f1-run-httpd\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.185367 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eeb4962a-1819-4562-baf3-1b27c476d5f1-log-httpd\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.188041 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eeb4962a-1819-4562-baf3-1b27c476d5f1-run-httpd\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.191495 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.194498 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.194696 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.197554 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-config-data\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.204858 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeb4962a-1819-4562-baf3-1b27c476d5f1-scripts\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.205987 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d8vc\" (UniqueName: \"kubernetes.io/projected/eeb4962a-1819-4562-baf3-1b27c476d5f1-kube-api-access-7d8vc\") pod \"ceilometer-0\" (UID: \"eeb4962a-1819-4562-baf3-1b27c476d5f1\") " pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.273310 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.279845 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.393237 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-catalog-content\") pod \"169ec2e6-4704-4425-9710-c5b4af3a609c\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.393448 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82skh\" (UniqueName: \"kubernetes.io/projected/169ec2e6-4704-4425-9710-c5b4af3a609c-kube-api-access-82skh\") pod \"169ec2e6-4704-4425-9710-c5b4af3a609c\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.393535 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-utilities\") pod \"169ec2e6-4704-4425-9710-c5b4af3a609c\" (UID: \"169ec2e6-4704-4425-9710-c5b4af3a609c\") " Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.396229 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-utilities" (OuterVolumeSpecName: "utilities") pod "169ec2e6-4704-4425-9710-c5b4af3a609c" (UID: "169ec2e6-4704-4425-9710-c5b4af3a609c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.407153 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/169ec2e6-4704-4425-9710-c5b4af3a609c-kube-api-access-82skh" (OuterVolumeSpecName: "kube-api-access-82skh") pod "169ec2e6-4704-4425-9710-c5b4af3a609c" (UID: "169ec2e6-4704-4425-9710-c5b4af3a609c"). InnerVolumeSpecName "kube-api-access-82skh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.490659 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "169ec2e6-4704-4425-9710-c5b4af3a609c" (UID: "169ec2e6-4704-4425-9710-c5b4af3a609c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.496794 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.496833 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82skh\" (UniqueName: \"kubernetes.io/projected/169ec2e6-4704-4425-9710-c5b4af3a609c-kube-api-access-82skh\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.496849 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/169ec2e6-4704-4425-9710-c5b4af3a609c-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.583120 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc78f25-c3cb-4627-be10-a6838c84d61f" path="/var/lib/kubelet/pods/afc78f25-c3cb-4627-be10-a6838c84d61f/volumes" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.761627 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.859689 4796 generic.go:334] "Generic (PLEG): container finished" podID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerID="19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2" exitCode=0 Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.859755 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m2mw5" event={"ID":"169ec2e6-4704-4425-9710-c5b4af3a609c","Type":"ContainerDied","Data":"19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2"} Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.859761 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m2mw5" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.859781 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m2mw5" event={"ID":"169ec2e6-4704-4425-9710-c5b4af3a609c","Type":"ContainerDied","Data":"b4e01eb29cee89872cc93acb0dc8c12bd7cff073b8fc6f59221a697815d004a1"} Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.859798 4796 scope.go:117] "RemoveContainer" containerID="19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.862692 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eeb4962a-1819-4562-baf3-1b27c476d5f1","Type":"ContainerStarted","Data":"1f08ae11cc8edc5851b4cb8a93970d92d2514a4fef6e9cdf73084a73ab47cd15"} Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.892985 4796 scope.go:117] "RemoveContainer" containerID="74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.898695 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m2mw5"] Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.908583 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m2mw5"] Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.914036 4796 scope.go:117] "RemoveContainer" containerID="005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.968068 4796 scope.go:117] "RemoveContainer" containerID="19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2" Nov 27 12:26:11 crc kubenswrapper[4796]: E1127 12:26:11.968725 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2\": container with ID starting with 19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2 not found: ID does not exist" containerID="19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.968773 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2"} err="failed to get container status \"19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2\": rpc error: code = NotFound desc = could not find container \"19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2\": container with ID starting with 19ab10833bf26c41f01414beb333c1008a7ceebe738c02747c0aa7308e6bd1d2 not found: ID does not exist" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.968792 4796 scope.go:117] "RemoveContainer" containerID="74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8" Nov 27 12:26:11 crc kubenswrapper[4796]: E1127 12:26:11.969058 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8\": container with ID starting with 74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8 not found: ID does not exist" containerID="74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.969099 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8"} err="failed to get container status \"74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8\": rpc error: code = NotFound desc = could not find container \"74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8\": container with ID starting with 74dc89786b17f64f7dc34952e925fc7e5514161608c9f89f24579394f92dd3d8 not found: ID does not exist" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.969129 4796 scope.go:117] "RemoveContainer" containerID="005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549" Nov 27 12:26:11 crc kubenswrapper[4796]: E1127 12:26:11.969465 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549\": container with ID starting with 005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549 not found: ID does not exist" containerID="005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549" Nov 27 12:26:11 crc kubenswrapper[4796]: I1127 12:26:11.969513 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549"} err="failed to get container status \"005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549\": rpc error: code = NotFound desc = could not find container \"005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549\": container with ID starting with 005609ba60921f9db7403b78d7fefeb2435a95f7a84b84d601f1a44800522549 not found: ID does not exist" Nov 27 12:26:12 crc kubenswrapper[4796]: I1127 12:26:12.873581 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eeb4962a-1819-4562-baf3-1b27c476d5f1","Type":"ContainerStarted","Data":"9b1bdfad222b481ff1f7a34f577045ffb6ec2b92b6e4e1a0ba9dc3c3deb9643b"} Nov 27 12:26:13 crc kubenswrapper[4796]: I1127 12:26:13.584868 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" path="/var/lib/kubelet/pods/169ec2e6-4704-4425-9710-c5b4af3a609c/volumes" Nov 27 12:26:13 crc kubenswrapper[4796]: I1127 12:26:13.884472 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eeb4962a-1819-4562-baf3-1b27c476d5f1","Type":"ContainerStarted","Data":"f9b7ddcda4f04e70a07c5c89cd14e367e80b9825a2665b5623ffda5fdbfc6c76"} Nov 27 12:26:14 crc kubenswrapper[4796]: I1127 12:26:14.894503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eeb4962a-1819-4562-baf3-1b27c476d5f1","Type":"ContainerStarted","Data":"1bd61d76850807f56ecae066c1d6f2e8790f437fd1540ef13438e377a1e9d2cb"} Nov 27 12:26:15 crc kubenswrapper[4796]: I1127 12:26:15.909998 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eeb4962a-1819-4562-baf3-1b27c476d5f1","Type":"ContainerStarted","Data":"e8db6aab0d8d59224db79b0547bfbb33cc94b7fcb7460a297cb3a08a9de5bc1e"} Nov 27 12:26:15 crc kubenswrapper[4796]: I1127 12:26:15.911631 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 12:26:15 crc kubenswrapper[4796]: I1127 12:26:15.940574 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.373203871 podStartE2EDuration="5.940551789s" podCreationTimestamp="2025-11-27 12:26:10 +0000 UTC" firstStartedPulling="2025-11-27 12:26:11.767512194 +0000 UTC m=+3689.285831112" lastFinishedPulling="2025-11-27 12:26:15.334860112 +0000 UTC m=+3692.853179030" observedRunningTime="2025-11-27 12:26:15.933436076 +0000 UTC m=+3693.451754994" watchObservedRunningTime="2025-11-27 12:26:15.940551789 +0000 UTC m=+3693.458870707" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.140189 4796 generic.go:334] "Generic (PLEG): container finished" podID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerID="18d27670bca6efcc8b66e4a40592b92bb831f2318ec4513ea5dc3dd828188684" exitCode=137 Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.140958 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerDied","Data":"18d27670bca6efcc8b66e4a40592b92bb831f2318ec4513ea5dc3dd828188684"} Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.337524 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.486499 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-config-data\") pod \"547e11ed-da55-4baa-b1f1-737907bd7ea1\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.486987 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-combined-ca-bundle\") pod \"547e11ed-da55-4baa-b1f1-737907bd7ea1\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.487117 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52zxd\" (UniqueName: \"kubernetes.io/projected/547e11ed-da55-4baa-b1f1-737907bd7ea1-kube-api-access-52zxd\") pod \"547e11ed-da55-4baa-b1f1-737907bd7ea1\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.487323 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-scripts\") pod \"547e11ed-da55-4baa-b1f1-737907bd7ea1\" (UID: \"547e11ed-da55-4baa-b1f1-737907bd7ea1\") " Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.493229 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547e11ed-da55-4baa-b1f1-737907bd7ea1-kube-api-access-52zxd" (OuterVolumeSpecName: "kube-api-access-52zxd") pod "547e11ed-da55-4baa-b1f1-737907bd7ea1" (UID: "547e11ed-da55-4baa-b1f1-737907bd7ea1"). InnerVolumeSpecName "kube-api-access-52zxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.493509 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-scripts" (OuterVolumeSpecName: "scripts") pod "547e11ed-da55-4baa-b1f1-737907bd7ea1" (UID: "547e11ed-da55-4baa-b1f1-737907bd7ea1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.589610 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.589646 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52zxd\" (UniqueName: \"kubernetes.io/projected/547e11ed-da55-4baa-b1f1-737907bd7ea1-kube-api-access-52zxd\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.595253 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "547e11ed-da55-4baa-b1f1-737907bd7ea1" (UID: "547e11ed-da55-4baa-b1f1-737907bd7ea1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.625247 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-config-data" (OuterVolumeSpecName: "config-data") pod "547e11ed-da55-4baa-b1f1-737907bd7ea1" (UID: "547e11ed-da55-4baa-b1f1-737907bd7ea1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.691059 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:40 crc kubenswrapper[4796]: I1127 12:26:40.695800 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547e11ed-da55-4baa-b1f1-737907bd7ea1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.154789 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"547e11ed-da55-4baa-b1f1-737907bd7ea1","Type":"ContainerDied","Data":"3067a3c162f82f285d1c3d794a09aacd83c7bdc60dbc6289ecef04378be99bbc"} Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.154861 4796 scope.go:117] "RemoveContainer" containerID="18d27670bca6efcc8b66e4a40592b92bb831f2318ec4513ea5dc3dd828188684" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.154886 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.186364 4796 scope.go:117] "RemoveContainer" containerID="09fe631702ec34e296b71b34910ee8356f8e025f59efbf35d00700d7b930a870" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.202305 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.211472 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.225093 4796 scope.go:117] "RemoveContainer" containerID="e12be2d2ee731c35bc837e77c016f7dc5bf36321c575b9261b1af298c2d16bac" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.234909 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235500 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-api" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235528 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-api" Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235541 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="extract-utilities" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235549 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="extract-utilities" Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235562 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="registry-server" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235570 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="registry-server" Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235589 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="extract-content" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235597 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="extract-content" Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235618 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-evaluator" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235626 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-evaluator" Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235643 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-listener" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235652 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-listener" Nov 27 12:26:41 crc kubenswrapper[4796]: E1127 12:26:41.235666 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-notifier" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235674 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-notifier" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235922 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-notifier" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235941 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-evaluator" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235955 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-listener" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235975 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="169ec2e6-4704-4425-9710-c5b4af3a609c" containerName="registry-server" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.235989 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" containerName="aodh-api" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.237940 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.244943 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.245500 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vs2ld" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.245854 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.246229 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.246308 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.246837 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.273553 4796 scope.go:117] "RemoveContainer" containerID="cf65a6d196417dd9e0d215320bb1ccbf54d8bf4a840fcfd1b73d13e55c5fff75" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.281317 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.306339 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-public-tls-certs\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.306486 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-scripts\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.306634 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-combined-ca-bundle\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.306712 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cng6p\" (UniqueName: \"kubernetes.io/projected/52f66cd2-21fb-4949-829f-1e05e6e010ee-kube-api-access-cng6p\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.306742 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-internal-tls-certs\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.306767 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-config-data\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.408928 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-public-tls-certs\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.408991 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-scripts\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.409059 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-combined-ca-bundle\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.409098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cng6p\" (UniqueName: \"kubernetes.io/projected/52f66cd2-21fb-4949-829f-1e05e6e010ee-kube-api-access-cng6p\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.409119 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-internal-tls-certs\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.409138 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-config-data\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.413325 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-scripts\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.413428 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-public-tls-certs\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.413751 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-config-data\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.413921 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-combined-ca-bundle\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.418113 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-internal-tls-certs\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.425367 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cng6p\" (UniqueName: \"kubernetes.io/projected/52f66cd2-21fb-4949-829f-1e05e6e010ee-kube-api-access-cng6p\") pod \"aodh-0\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.562699 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:26:41 crc kubenswrapper[4796]: I1127 12:26:41.596802 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="547e11ed-da55-4baa-b1f1-737907bd7ea1" path="/var/lib/kubelet/pods/547e11ed-da55-4baa-b1f1-737907bd7ea1/volumes" Nov 27 12:26:42 crc kubenswrapper[4796]: I1127 12:26:42.031927 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:26:42 crc kubenswrapper[4796]: W1127 12:26:42.037156 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice/crio-396b3e4e378dfe25701126bc0748cee8c8e621fdcd8981f111fdfae6a86d06ab WatchSource:0}: Error finding container 396b3e4e378dfe25701126bc0748cee8c8e621fdcd8981f111fdfae6a86d06ab: Status 404 returned error can't find the container with id 396b3e4e378dfe25701126bc0748cee8c8e621fdcd8981f111fdfae6a86d06ab Nov 27 12:26:42 crc kubenswrapper[4796]: I1127 12:26:42.170227 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerStarted","Data":"396b3e4e378dfe25701126bc0748cee8c8e621fdcd8981f111fdfae6a86d06ab"} Nov 27 12:26:43 crc kubenswrapper[4796]: I1127 12:26:43.179446 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerStarted","Data":"5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb"} Nov 27 12:26:44 crc kubenswrapper[4796]: I1127 12:26:44.190373 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerStarted","Data":"6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5"} Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.205043 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerStarted","Data":"1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628"} Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.911289 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7l762"] Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.914103 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.919616 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-catalog-content\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.919744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8fkh\" (UniqueName: \"kubernetes.io/projected/4d8dfd55-eca2-4d9f-9d58-818599f6c280-kube-api-access-h8fkh\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.919928 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-utilities\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:45 crc kubenswrapper[4796]: I1127 12:26:45.927036 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7l762"] Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.021840 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-utilities\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.022202 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-catalog-content\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.022224 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8fkh\" (UniqueName: \"kubernetes.io/projected/4d8dfd55-eca2-4d9f-9d58-818599f6c280-kube-api-access-h8fkh\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.022505 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-utilities\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.022852 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-catalog-content\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.046204 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8fkh\" (UniqueName: \"kubernetes.io/projected/4d8dfd55-eca2-4d9f-9d58-818599f6c280-kube-api-access-h8fkh\") pod \"redhat-marketplace-7l762\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.216327 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerStarted","Data":"97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342"} Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.234686 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.252461 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.139078215 podStartE2EDuration="5.252430493s" podCreationTimestamp="2025-11-27 12:26:41 +0000 UTC" firstStartedPulling="2025-11-27 12:26:42.041495486 +0000 UTC m=+3719.559814404" lastFinishedPulling="2025-11-27 12:26:45.154847764 +0000 UTC m=+3722.673166682" observedRunningTime="2025-11-27 12:26:46.238259938 +0000 UTC m=+3723.756578856" watchObservedRunningTime="2025-11-27 12:26:46.252430493 +0000 UTC m=+3723.770749411" Nov 27 12:26:46 crc kubenswrapper[4796]: I1127 12:26:46.798789 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7l762"] Nov 27 12:26:47 crc kubenswrapper[4796]: I1127 12:26:47.231461 4796 generic.go:334] "Generic (PLEG): container finished" podID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerID="b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737" exitCode=0 Nov 27 12:26:47 crc kubenswrapper[4796]: I1127 12:26:47.231634 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7l762" event={"ID":"4d8dfd55-eca2-4d9f-9d58-818599f6c280","Type":"ContainerDied","Data":"b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737"} Nov 27 12:26:47 crc kubenswrapper[4796]: I1127 12:26:47.231866 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7l762" event={"ID":"4d8dfd55-eca2-4d9f-9d58-818599f6c280","Type":"ContainerStarted","Data":"909e6e99ecad500e6f74a7568de1bf54e870e63416b89dd649bfa779fec81330"} Nov 27 12:26:49 crc kubenswrapper[4796]: I1127 12:26:49.250780 4796 generic.go:334] "Generic (PLEG): container finished" podID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerID="de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e" exitCode=0 Nov 27 12:26:49 crc kubenswrapper[4796]: I1127 12:26:49.250996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7l762" event={"ID":"4d8dfd55-eca2-4d9f-9d58-818599f6c280","Type":"ContainerDied","Data":"de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e"} Nov 27 12:26:50 crc kubenswrapper[4796]: I1127 12:26:50.263905 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7l762" event={"ID":"4d8dfd55-eca2-4d9f-9d58-818599f6c280","Type":"ContainerStarted","Data":"a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93"} Nov 27 12:26:50 crc kubenswrapper[4796]: I1127 12:26:50.287628 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7l762" podStartSLOduration=2.793509388 podStartE2EDuration="5.287610379s" podCreationTimestamp="2025-11-27 12:26:45 +0000 UTC" firstStartedPulling="2025-11-27 12:26:47.233683518 +0000 UTC m=+3724.752002436" lastFinishedPulling="2025-11-27 12:26:49.727784509 +0000 UTC m=+3727.246103427" observedRunningTime="2025-11-27 12:26:50.282380306 +0000 UTC m=+3727.800699234" watchObservedRunningTime="2025-11-27 12:26:50.287610379 +0000 UTC m=+3727.805929297" Nov 27 12:26:56 crc kubenswrapper[4796]: I1127 12:26:56.236110 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:56 crc kubenswrapper[4796]: I1127 12:26:56.236678 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:56 crc kubenswrapper[4796]: I1127 12:26:56.279865 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:56 crc kubenswrapper[4796]: I1127 12:26:56.361698 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:56 crc kubenswrapper[4796]: I1127 12:26:56.531247 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7l762"] Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.338854 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7l762" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="registry-server" containerID="cri-o://a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93" gracePeriod=2 Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.792130 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.868259 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-utilities\") pod \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.868632 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-catalog-content\") pod \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.868823 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8fkh\" (UniqueName: \"kubernetes.io/projected/4d8dfd55-eca2-4d9f-9d58-818599f6c280-kube-api-access-h8fkh\") pod \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\" (UID: \"4d8dfd55-eca2-4d9f-9d58-818599f6c280\") " Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.869675 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-utilities" (OuterVolumeSpecName: "utilities") pod "4d8dfd55-eca2-4d9f-9d58-818599f6c280" (UID: "4d8dfd55-eca2-4d9f-9d58-818599f6c280"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.874688 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d8dfd55-eca2-4d9f-9d58-818599f6c280-kube-api-access-h8fkh" (OuterVolumeSpecName: "kube-api-access-h8fkh") pod "4d8dfd55-eca2-4d9f-9d58-818599f6c280" (UID: "4d8dfd55-eca2-4d9f-9d58-818599f6c280"). InnerVolumeSpecName "kube-api-access-h8fkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.889047 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d8dfd55-eca2-4d9f-9d58-818599f6c280" (UID: "4d8dfd55-eca2-4d9f-9d58-818599f6c280"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.971324 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8fkh\" (UniqueName: \"kubernetes.io/projected/4d8dfd55-eca2-4d9f-9d58-818599f6c280-kube-api-access-h8fkh\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.971366 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:58 crc kubenswrapper[4796]: I1127 12:26:58.971376 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8dfd55-eca2-4d9f-9d58-818599f6c280-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.350937 4796 generic.go:334] "Generic (PLEG): container finished" podID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerID="a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93" exitCode=0 Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.351010 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7l762" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.351046 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7l762" event={"ID":"4d8dfd55-eca2-4d9f-9d58-818599f6c280","Type":"ContainerDied","Data":"a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93"} Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.351327 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7l762" event={"ID":"4d8dfd55-eca2-4d9f-9d58-818599f6c280","Type":"ContainerDied","Data":"909e6e99ecad500e6f74a7568de1bf54e870e63416b89dd649bfa779fec81330"} Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.351377 4796 scope.go:117] "RemoveContainer" containerID="a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.372027 4796 scope.go:117] "RemoveContainer" containerID="de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.390521 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7l762"] Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.397228 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7l762"] Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.399115 4796 scope.go:117] "RemoveContainer" containerID="b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.462347 4796 scope.go:117] "RemoveContainer" containerID="a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93" Nov 27 12:26:59 crc kubenswrapper[4796]: E1127 12:26:59.462826 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93\": container with ID starting with a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93 not found: ID does not exist" containerID="a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.462866 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93"} err="failed to get container status \"a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93\": rpc error: code = NotFound desc = could not find container \"a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93\": container with ID starting with a0f45d09e3b8fa54e555ae8c6e8e0e2cbd05df28940c9522ec4110d8d08b6c93 not found: ID does not exist" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.462889 4796 scope.go:117] "RemoveContainer" containerID="de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e" Nov 27 12:26:59 crc kubenswrapper[4796]: E1127 12:26:59.463208 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e\": container with ID starting with de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e not found: ID does not exist" containerID="de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.463232 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e"} err="failed to get container status \"de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e\": rpc error: code = NotFound desc = could not find container \"de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e\": container with ID starting with de51eef0d9553b6adb2a39aa9f11d79c34e4efb20c4192b3965b8c67fc56cd8e not found: ID does not exist" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.463246 4796 scope.go:117] "RemoveContainer" containerID="b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737" Nov 27 12:26:59 crc kubenswrapper[4796]: E1127 12:26:59.463578 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737\": container with ID starting with b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737 not found: ID does not exist" containerID="b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.463615 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737"} err="failed to get container status \"b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737\": rpc error: code = NotFound desc = could not find container \"b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737\": container with ID starting with b4d2851d472e4af9beb068f6f4f1e636528814f390db6d9a085523964a595737 not found: ID does not exist" Nov 27 12:26:59 crc kubenswrapper[4796]: I1127 12:26:59.579425 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" path="/var/lib/kubelet/pods/4d8dfd55-eca2-4d9f-9d58-818599f6c280/volumes" Nov 27 12:28:31 crc kubenswrapper[4796]: I1127 12:28:31.879312 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:28:31 crc kubenswrapper[4796]: I1127 12:28:31.880754 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:29:01 crc kubenswrapper[4796]: I1127 12:29:01.879258 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:29:01 crc kubenswrapper[4796]: I1127 12:29:01.880021 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:29:31 crc kubenswrapper[4796]: I1127 12:29:31.879385 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:29:31 crc kubenswrapper[4796]: I1127 12:29:31.879841 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:29:31 crc kubenswrapper[4796]: I1127 12:29:31.879886 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:29:31 crc kubenswrapper[4796]: I1127 12:29:31.880666 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:29:31 crc kubenswrapper[4796]: I1127 12:29:31.880717 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" gracePeriod=600 Nov 27 12:29:32 crc kubenswrapper[4796]: E1127 12:29:32.002507 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:29:32 crc kubenswrapper[4796]: I1127 12:29:32.735672 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" exitCode=0 Nov 27 12:29:32 crc kubenswrapper[4796]: I1127 12:29:32.735738 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58"} Nov 27 12:29:32 crc kubenswrapper[4796]: I1127 12:29:32.736005 4796 scope.go:117] "RemoveContainer" containerID="9b9a3460a4a4c54e56ba38ba7e96ca41227b4bbafda0b43cac4ec6cea53fcf07" Nov 27 12:29:32 crc kubenswrapper[4796]: I1127 12:29:32.736623 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:29:32 crc kubenswrapper[4796]: E1127 12:29:32.736867 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:29:39 crc kubenswrapper[4796]: I1127 12:29:39.861242 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:29:46 crc kubenswrapper[4796]: I1127 12:29:46.569335 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:29:46 crc kubenswrapper[4796]: E1127 12:29:46.569903 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.488128 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp"] Nov 27 12:29:53 crc kubenswrapper[4796]: E1127 12:29:53.489993 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="extract-utilities" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.490074 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="extract-utilities" Nov 27 12:29:53 crc kubenswrapper[4796]: E1127 12:29:53.490149 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="extract-content" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.490207 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="extract-content" Nov 27 12:29:53 crc kubenswrapper[4796]: E1127 12:29:53.490320 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="registry-server" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.490390 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="registry-server" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.490652 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d8dfd55-eca2-4d9f-9d58-818599f6c280" containerName="registry-server" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.492171 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.496100 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.496750 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp"] Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.599658 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.600008 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgbd\" (UniqueName: \"kubernetes.io/projected/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-kube-api-access-czgbd\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.600143 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.702241 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.702324 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czgbd\" (UniqueName: \"kubernetes.io/projected/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-kube-api-access-czgbd\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.702445 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.702899 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.703609 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.720502 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgbd\" (UniqueName: \"kubernetes.io/projected/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-kube-api-access-czgbd\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:53 crc kubenswrapper[4796]: I1127 12:29:53.816208 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:54 crc kubenswrapper[4796]: I1127 12:29:54.240107 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp"] Nov 27 12:29:54 crc kubenswrapper[4796]: I1127 12:29:54.943207 4796 generic.go:334] "Generic (PLEG): container finished" podID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerID="84613cb69fa468c7551f25506f23f09b7a168fbc8dd2659bd76131387b2fa5b9" exitCode=0 Nov 27 12:29:54 crc kubenswrapper[4796]: I1127 12:29:54.943330 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" event={"ID":"076bc7ca-c3ef-4c76-b896-d4bdac2191e4","Type":"ContainerDied","Data":"84613cb69fa468c7551f25506f23f09b7a168fbc8dd2659bd76131387b2fa5b9"} Nov 27 12:29:54 crc kubenswrapper[4796]: I1127 12:29:54.943520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" event={"ID":"076bc7ca-c3ef-4c76-b896-d4bdac2191e4","Type":"ContainerStarted","Data":"4aa29e2f67cae1769021dc1722e08b32fd412251ea822bc5afa59d142bdd943c"} Nov 27 12:29:54 crc kubenswrapper[4796]: I1127 12:29:54.945615 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:29:56 crc kubenswrapper[4796]: I1127 12:29:56.989152 4796 generic.go:334] "Generic (PLEG): container finished" podID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerID="02e5d1de963c53778fd907882be1a062f7b837bd726e7eadcae35b6aeba0f5af" exitCode=0 Nov 27 12:29:56 crc kubenswrapper[4796]: I1127 12:29:56.989241 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" event={"ID":"076bc7ca-c3ef-4c76-b896-d4bdac2191e4","Type":"ContainerDied","Data":"02e5d1de963c53778fd907882be1a062f7b837bd726e7eadcae35b6aeba0f5af"} Nov 27 12:29:58 crc kubenswrapper[4796]: I1127 12:29:58.007782 4796 generic.go:334] "Generic (PLEG): container finished" podID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerID="f11bf5558063cb410858517b2f61b28105e1cb0e391c4a88e5b7d85eebfd438d" exitCode=0 Nov 27 12:29:58 crc kubenswrapper[4796]: I1127 12:29:58.008010 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" event={"ID":"076bc7ca-c3ef-4c76-b896-d4bdac2191e4","Type":"ContainerDied","Data":"f11bf5558063cb410858517b2f61b28105e1cb0e391c4a88e5b7d85eebfd438d"} Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.342428 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.512528 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-util\") pod \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.512623 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czgbd\" (UniqueName: \"kubernetes.io/projected/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-kube-api-access-czgbd\") pod \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.512751 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-bundle\") pod \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\" (UID: \"076bc7ca-c3ef-4c76-b896-d4bdac2191e4\") " Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.515763 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-bundle" (OuterVolumeSpecName: "bundle") pod "076bc7ca-c3ef-4c76-b896-d4bdac2191e4" (UID: "076bc7ca-c3ef-4c76-b896-d4bdac2191e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.519455 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-kube-api-access-czgbd" (OuterVolumeSpecName: "kube-api-access-czgbd") pod "076bc7ca-c3ef-4c76-b896-d4bdac2191e4" (UID: "076bc7ca-c3ef-4c76-b896-d4bdac2191e4"). InnerVolumeSpecName "kube-api-access-czgbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.531727 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-util" (OuterVolumeSpecName: "util") pod "076bc7ca-c3ef-4c76-b896-d4bdac2191e4" (UID: "076bc7ca-c3ef-4c76-b896-d4bdac2191e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.570671 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:29:59 crc kubenswrapper[4796]: E1127 12:29:59.570901 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.614549 4796 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.614574 4796 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-util\") on node \"crc\" DevicePath \"\"" Nov 27 12:29:59 crc kubenswrapper[4796]: I1127 12:29:59.614584 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czgbd\" (UniqueName: \"kubernetes.io/projected/076bc7ca-c3ef-4c76-b896-d4bdac2191e4-kube-api-access-czgbd\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.043936 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" event={"ID":"076bc7ca-c3ef-4c76-b896-d4bdac2191e4","Type":"ContainerDied","Data":"4aa29e2f67cae1769021dc1722e08b32fd412251ea822bc5afa59d142bdd943c"} Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.044008 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4aa29e2f67cae1769021dc1722e08b32fd412251ea822bc5afa59d142bdd943c" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.044010 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.177735 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl"] Nov 27 12:30:00 crc kubenswrapper[4796]: E1127 12:30:00.178189 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="pull" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.178209 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="pull" Nov 27 12:30:00 crc kubenswrapper[4796]: E1127 12:30:00.178252 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="util" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.178258 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="util" Nov 27 12:30:00 crc kubenswrapper[4796]: E1127 12:30:00.178284 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="extract" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.178290 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="extract" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.178485 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="076bc7ca-c3ef-4c76-b896-d4bdac2191e4" containerName="extract" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.179159 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.181821 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.186260 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.200152 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl"] Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.329726 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-secret-volume\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.329852 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpttk\" (UniqueName: \"kubernetes.io/projected/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-kube-api-access-xpttk\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.330165 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-config-volume\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.432198 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-secret-volume\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.432260 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpttk\" (UniqueName: \"kubernetes.io/projected/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-kube-api-access-xpttk\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.432374 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-config-volume\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.433135 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-config-volume\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.437661 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-secret-volume\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.448030 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpttk\" (UniqueName: \"kubernetes.io/projected/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-kube-api-access-xpttk\") pod \"collect-profiles-29404110-h4dpl\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.502831 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:00 crc kubenswrapper[4796]: I1127 12:30:00.971991 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl"] Nov 27 12:30:00 crc kubenswrapper[4796]: W1127 12:30:00.977454 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9ef66fc_10ce_4734_af9a_ebfb6e678f48.slice/crio-7c5052bcb7193b1b6fa12137ade00ba71b7d1efb9e36b2ebe6b00c37926a8f05 WatchSource:0}: Error finding container 7c5052bcb7193b1b6fa12137ade00ba71b7d1efb9e36b2ebe6b00c37926a8f05: Status 404 returned error can't find the container with id 7c5052bcb7193b1b6fa12137ade00ba71b7d1efb9e36b2ebe6b00c37926a8f05 Nov 27 12:30:01 crc kubenswrapper[4796]: I1127 12:30:01.054316 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" event={"ID":"b9ef66fc-10ce-4734-af9a-ebfb6e678f48","Type":"ContainerStarted","Data":"7c5052bcb7193b1b6fa12137ade00ba71b7d1efb9e36b2ebe6b00c37926a8f05"} Nov 27 12:30:02 crc kubenswrapper[4796]: I1127 12:30:02.066171 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" event={"ID":"b9ef66fc-10ce-4734-af9a-ebfb6e678f48","Type":"ContainerStarted","Data":"3176793b4bba5f00b2a3b4da17b723bb7d4b10db06ccd2311440ea8672f1da0e"} Nov 27 12:30:02 crc kubenswrapper[4796]: I1127 12:30:02.088346 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" podStartSLOduration=2.088310481 podStartE2EDuration="2.088310481s" podCreationTimestamp="2025-11-27 12:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:30:02.080386506 +0000 UTC m=+3919.598705424" watchObservedRunningTime="2025-11-27 12:30:02.088310481 +0000 UTC m=+3919.606629399" Nov 27 12:30:03 crc kubenswrapper[4796]: I1127 12:30:03.077476 4796 generic.go:334] "Generic (PLEG): container finished" podID="b9ef66fc-10ce-4734-af9a-ebfb6e678f48" containerID="3176793b4bba5f00b2a3b4da17b723bb7d4b10db06ccd2311440ea8672f1da0e" exitCode=0 Nov 27 12:30:03 crc kubenswrapper[4796]: I1127 12:30:03.077573 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" event={"ID":"b9ef66fc-10ce-4734-af9a-ebfb6e678f48","Type":"ContainerDied","Data":"3176793b4bba5f00b2a3b4da17b723bb7d4b10db06ccd2311440ea8672f1da0e"} Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.415720 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.510029 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-config-volume\") pod \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.510128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-secret-volume\") pod \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.510152 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpttk\" (UniqueName: \"kubernetes.io/projected/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-kube-api-access-xpttk\") pod \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\" (UID: \"b9ef66fc-10ce-4734-af9a-ebfb6e678f48\") " Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.511462 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9ef66fc-10ce-4734-af9a-ebfb6e678f48" (UID: "b9ef66fc-10ce-4734-af9a-ebfb6e678f48"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.515950 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-kube-api-access-xpttk" (OuterVolumeSpecName: "kube-api-access-xpttk") pod "b9ef66fc-10ce-4734-af9a-ebfb6e678f48" (UID: "b9ef66fc-10ce-4734-af9a-ebfb6e678f48"). InnerVolumeSpecName "kube-api-access-xpttk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.516151 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9ef66fc-10ce-4734-af9a-ebfb6e678f48" (UID: "b9ef66fc-10ce-4734-af9a-ebfb6e678f48"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.612593 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.612626 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:04 crc kubenswrapper[4796]: I1127 12:30:04.612636 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpttk\" (UniqueName: \"kubernetes.io/projected/b9ef66fc-10ce-4734-af9a-ebfb6e678f48-kube-api-access-xpttk\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:05 crc kubenswrapper[4796]: I1127 12:30:05.124743 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" event={"ID":"b9ef66fc-10ce-4734-af9a-ebfb6e678f48","Type":"ContainerDied","Data":"7c5052bcb7193b1b6fa12137ade00ba71b7d1efb9e36b2ebe6b00c37926a8f05"} Nov 27 12:30:05 crc kubenswrapper[4796]: I1127 12:30:05.125033 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c5052bcb7193b1b6fa12137ade00ba71b7d1efb9e36b2ebe6b00c37926a8f05" Nov 27 12:30:05 crc kubenswrapper[4796]: I1127 12:30:05.125083 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404110-h4dpl" Nov 27 12:30:05 crc kubenswrapper[4796]: I1127 12:30:05.193784 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z"] Nov 27 12:30:05 crc kubenswrapper[4796]: I1127 12:30:05.204874 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404065-kfm5z"] Nov 27 12:30:05 crc kubenswrapper[4796]: I1127 12:30:05.598358 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b06300-6722-4ac8-b033-4781f968e40f" path="/var/lib/kubelet/pods/b2b06300-6722-4ac8-b033-4781f968e40f/volumes" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.825740 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9"] Nov 27 12:30:11 crc kubenswrapper[4796]: E1127 12:30:11.826845 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ef66fc-10ce-4734-af9a-ebfb6e678f48" containerName="collect-profiles" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.826863 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ef66fc-10ce-4734-af9a-ebfb6e678f48" containerName="collect-profiles" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.827114 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ef66fc-10ce-4734-af9a-ebfb6e678f48" containerName="collect-profiles" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.827985 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.831289 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-c2hx6" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.831526 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.837547 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.845577 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622"] Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.847161 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.850301 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-kmv2c" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.850687 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.856261 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9"] Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.868995 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622"] Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.883486 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw"] Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.884807 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.903932 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw"] Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.972739 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd9f4f30-7f8c-40d0-b559-368b90f90b4b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-n5622\" (UID: \"bd9f4f30-7f8c-40d0-b559-368b90f90b4b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.972945 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx6xw\" (UniqueName: \"kubernetes.io/projected/12aebb1f-5956-4364-b2c9-80cf77ddb77a-kube-api-access-tx6xw\") pod \"obo-prometheus-operator-668cf9dfbb-q7hl9\" (UID: \"12aebb1f-5956-4364-b2c9-80cf77ddb77a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" Nov 27 12:30:11 crc kubenswrapper[4796]: I1127 12:30:11.973288 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd9f4f30-7f8c-40d0-b559-368b90f90b4b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-n5622\" (UID: \"bd9f4f30-7f8c-40d0-b559-368b90f90b4b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.007769 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-x26pv"] Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.009334 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.012902 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.013290 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-pvfpd" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.025450 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-x26pv"] Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.075624 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e2638406-4d7c-404f-9906-544ce502974b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-bbzgw\" (UID: \"e2638406-4d7c-404f-9906-544ce502974b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.075970 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd9f4f30-7f8c-40d0-b559-368b90f90b4b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-n5622\" (UID: \"bd9f4f30-7f8c-40d0-b559-368b90f90b4b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.076104 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e2638406-4d7c-404f-9906-544ce502974b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-bbzgw\" (UID: \"e2638406-4d7c-404f-9906-544ce502974b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.076220 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx6xw\" (UniqueName: \"kubernetes.io/projected/12aebb1f-5956-4364-b2c9-80cf77ddb77a-kube-api-access-tx6xw\") pod \"obo-prometheus-operator-668cf9dfbb-q7hl9\" (UID: \"12aebb1f-5956-4364-b2c9-80cf77ddb77a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.076504 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd9f4f30-7f8c-40d0-b559-368b90f90b4b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-n5622\" (UID: \"bd9f4f30-7f8c-40d0-b559-368b90f90b4b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.086543 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd9f4f30-7f8c-40d0-b559-368b90f90b4b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-n5622\" (UID: \"bd9f4f30-7f8c-40d0-b559-368b90f90b4b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.086628 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd9f4f30-7f8c-40d0-b559-368b90f90b4b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-n5622\" (UID: \"bd9f4f30-7f8c-40d0-b559-368b90f90b4b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.098836 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx6xw\" (UniqueName: \"kubernetes.io/projected/12aebb1f-5956-4364-b2c9-80cf77ddb77a-kube-api-access-tx6xw\") pod \"obo-prometheus-operator-668cf9dfbb-q7hl9\" (UID: \"12aebb1f-5956-4364-b2c9-80cf77ddb77a\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.125657 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-t8f4v"] Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.127337 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.130471 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-rbwnx" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.140935 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-t8f4v"] Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.151827 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.175281 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.178464 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e2638406-4d7c-404f-9906-544ce502974b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-bbzgw\" (UID: \"e2638406-4d7c-404f-9906-544ce502974b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.178595 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e2638406-4d7c-404f-9906-544ce502974b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-bbzgw\" (UID: \"e2638406-4d7c-404f-9906-544ce502974b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.178699 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b928049-ffbb-4f1a-9d05-14f4bae23427-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-x26pv\" (UID: \"2b928049-ffbb-4f1a-9d05-14f4bae23427\") " pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.178732 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkpzr\" (UniqueName: \"kubernetes.io/projected/2b928049-ffbb-4f1a-9d05-14f4bae23427-kube-api-access-qkpzr\") pod \"observability-operator-d8bb48f5d-x26pv\" (UID: \"2b928049-ffbb-4f1a-9d05-14f4bae23427\") " pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.186808 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e2638406-4d7c-404f-9906-544ce502974b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-bbzgw\" (UID: \"e2638406-4d7c-404f-9906-544ce502974b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.186919 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e2638406-4d7c-404f-9906-544ce502974b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67d796579b-bbzgw\" (UID: \"e2638406-4d7c-404f-9906-544ce502974b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.200903 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.301875 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/94471100-b73d-4ca3-be61-4eac947eae64-openshift-service-ca\") pod \"perses-operator-5446b9c989-t8f4v\" (UID: \"94471100-b73d-4ca3-be61-4eac947eae64\") " pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.302104 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kj4c\" (UniqueName: \"kubernetes.io/projected/94471100-b73d-4ca3-be61-4eac947eae64-kube-api-access-4kj4c\") pod \"perses-operator-5446b9c989-t8f4v\" (UID: \"94471100-b73d-4ca3-be61-4eac947eae64\") " pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.302458 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b928049-ffbb-4f1a-9d05-14f4bae23427-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-x26pv\" (UID: \"2b928049-ffbb-4f1a-9d05-14f4bae23427\") " pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.302484 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkpzr\" (UniqueName: \"kubernetes.io/projected/2b928049-ffbb-4f1a-9d05-14f4bae23427-kube-api-access-qkpzr\") pod \"observability-operator-d8bb48f5d-x26pv\" (UID: \"2b928049-ffbb-4f1a-9d05-14f4bae23427\") " pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.342188 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b928049-ffbb-4f1a-9d05-14f4bae23427-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-x26pv\" (UID: \"2b928049-ffbb-4f1a-9d05-14f4bae23427\") " pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.403139 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkpzr\" (UniqueName: \"kubernetes.io/projected/2b928049-ffbb-4f1a-9d05-14f4bae23427-kube-api-access-qkpzr\") pod \"observability-operator-d8bb48f5d-x26pv\" (UID: \"2b928049-ffbb-4f1a-9d05-14f4bae23427\") " pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.427791 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/94471100-b73d-4ca3-be61-4eac947eae64-openshift-service-ca\") pod \"perses-operator-5446b9c989-t8f4v\" (UID: \"94471100-b73d-4ca3-be61-4eac947eae64\") " pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.427843 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kj4c\" (UniqueName: \"kubernetes.io/projected/94471100-b73d-4ca3-be61-4eac947eae64-kube-api-access-4kj4c\") pod \"perses-operator-5446b9c989-t8f4v\" (UID: \"94471100-b73d-4ca3-be61-4eac947eae64\") " pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.429450 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/94471100-b73d-4ca3-be61-4eac947eae64-openshift-service-ca\") pod \"perses-operator-5446b9c989-t8f4v\" (UID: \"94471100-b73d-4ca3-be61-4eac947eae64\") " pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.466677 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kj4c\" (UniqueName: \"kubernetes.io/projected/94471100-b73d-4ca3-be61-4eac947eae64-kube-api-access-4kj4c\") pod \"perses-operator-5446b9c989-t8f4v\" (UID: \"94471100-b73d-4ca3-be61-4eac947eae64\") " pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.573287 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:30:12 crc kubenswrapper[4796]: E1127 12:30:12.573877 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.631063 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:12 crc kubenswrapper[4796]: I1127 12:30:12.672747 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.056502 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622"] Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.165480 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw"] Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.204810 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" event={"ID":"bd9f4f30-7f8c-40d0-b559-368b90f90b4b","Type":"ContainerStarted","Data":"f220f914c71f8be8c193707a41e482323fa83f05f294a61676cb4f4f204e77a3"} Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.207520 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" event={"ID":"e2638406-4d7c-404f-9906-544ce502974b","Type":"ContainerStarted","Data":"54812a84328fb391fee85615364dbf910ff7e2a8285dc488f55d19d5dd191c89"} Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.209764 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9"] Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.288206 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-x26pv"] Nov 27 12:30:13 crc kubenswrapper[4796]: I1127 12:30:13.300818 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-t8f4v"] Nov 27 12:30:14 crc kubenswrapper[4796]: I1127 12:30:14.216844 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" event={"ID":"12aebb1f-5956-4364-b2c9-80cf77ddb77a","Type":"ContainerStarted","Data":"662db1f05e943a7e268b72fdfd73581c62ed8e14729e008d8e010d66cda75c01"} Nov 27 12:30:14 crc kubenswrapper[4796]: I1127 12:30:14.218403 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" event={"ID":"2b928049-ffbb-4f1a-9d05-14f4bae23427","Type":"ContainerStarted","Data":"dc3ef7fa15a7898e5da4e1a2980fe335fe3b4316e14c219b27bc74e8a96a956e"} Nov 27 12:30:14 crc kubenswrapper[4796]: I1127 12:30:14.219795 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" event={"ID":"94471100-b73d-4ca3-be61-4eac947eae64","Type":"ContainerStarted","Data":"da325c30492d374c9a2aa72d0b8e94a3b5119186239e14dfd18f60d41b026243"} Nov 27 12:30:26 crc kubenswrapper[4796]: I1127 12:30:26.569876 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:30:26 crc kubenswrapper[4796]: E1127 12:30:26.570727 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:30:28 crc kubenswrapper[4796]: E1127 12:30:28.998504 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3" Nov 27 12:30:29 crc kubenswrapper[4796]: E1127 12:30:28.999131 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tx6xw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-668cf9dfbb-q7hl9_openshift-operators(12aebb1f-5956-4364-b2c9-80cf77ddb77a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 12:30:29 crc kubenswrapper[4796]: E1127 12:30:29.000427 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" podUID="12aebb1f-5956-4364-b2c9-80cf77ddb77a" Nov 27 12:30:29 crc kubenswrapper[4796]: E1127 12:30:29.033377 4796 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Nov 27 12:30:29 crc kubenswrapper[4796]: E1127 12:30:29.033555 4796 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-67d796579b-bbzgw_openshift-operators(e2638406-4d7c-404f-9906-544ce502974b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 12:30:29 crc kubenswrapper[4796]: E1127 12:30:29.034738 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" podUID="e2638406-4d7c-404f-9906-544ce502974b" Nov 27 12:30:29 crc kubenswrapper[4796]: E1127 12:30:29.425581 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3\\\"\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" podUID="12aebb1f-5956-4364-b2c9-80cf77ddb77a" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.434909 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" event={"ID":"94471100-b73d-4ca3-be61-4eac947eae64","Type":"ContainerStarted","Data":"273593d51c938effc7690d2c9a1503cedcfa77be66b9f9f854d39339e691d3f3"} Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.435699 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.436796 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" event={"ID":"bd9f4f30-7f8c-40d0-b559-368b90f90b4b","Type":"ContainerStarted","Data":"57f023db07172ae86912cb61627c58ed9410a4fdf5e703bc1ec49ec6c6ccc75e"} Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.439380 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" event={"ID":"e2638406-4d7c-404f-9906-544ce502974b","Type":"ContainerStarted","Data":"44eb8c6b4a298f622be3be5a44a2a4176ee4684f66ddcb3534e84ced89292aae"} Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.440911 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" event={"ID":"2b928049-ffbb-4f1a-9d05-14f4bae23427","Type":"ContainerStarted","Data":"4243db9e8409d476b686c3d5ac91299e135b7b7e73c4ff3bd0b8128f955a2859"} Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.441238 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.458093 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" podStartSLOduration=2.7096223090000002 podStartE2EDuration="18.45807244s" podCreationTimestamp="2025-11-27 12:30:12 +0000 UTC" firstStartedPulling="2025-11-27 12:30:13.298045899 +0000 UTC m=+3930.816364817" lastFinishedPulling="2025-11-27 12:30:29.04649601 +0000 UTC m=+3946.564814948" observedRunningTime="2025-11-27 12:30:30.450461194 +0000 UTC m=+3947.968780112" watchObservedRunningTime="2025-11-27 12:30:30.45807244 +0000 UTC m=+3947.976391378" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.468221 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.468309 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-n5622" podStartSLOduration=3.4843113900000002 podStartE2EDuration="19.468291098s" podCreationTimestamp="2025-11-27 12:30:11 +0000 UTC" firstStartedPulling="2025-11-27 12:30:13.059089798 +0000 UTC m=+3930.577408716" lastFinishedPulling="2025-11-27 12:30:29.043069496 +0000 UTC m=+3946.561388424" observedRunningTime="2025-11-27 12:30:30.465187884 +0000 UTC m=+3947.983506812" watchObservedRunningTime="2025-11-27 12:30:30.468291098 +0000 UTC m=+3947.986610006" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.499214 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67d796579b-bbzgw" podStartSLOduration=-9223372017.355585 podStartE2EDuration="19.499191339s" podCreationTimestamp="2025-11-27 12:30:11 +0000 UTC" firstStartedPulling="2025-11-27 12:30:13.168929606 +0000 UTC m=+3930.687248534" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:30:30.494250814 +0000 UTC m=+3948.012569732" watchObservedRunningTime="2025-11-27 12:30:30.499191339 +0000 UTC m=+3948.017510257" Nov 27 12:30:30 crc kubenswrapper[4796]: I1127 12:30:30.543323 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-x26pv" podStartSLOduration=3.786489652 podStartE2EDuration="19.543303549s" podCreationTimestamp="2025-11-27 12:30:11 +0000 UTC" firstStartedPulling="2025-11-27 12:30:13.306082428 +0000 UTC m=+3930.824401346" lastFinishedPulling="2025-11-27 12:30:29.062896315 +0000 UTC m=+3946.581215243" observedRunningTime="2025-11-27 12:30:30.523937902 +0000 UTC m=+3948.042256820" watchObservedRunningTime="2025-11-27 12:30:30.543303549 +0000 UTC m=+3948.061622467" Nov 27 12:30:38 crc kubenswrapper[4796]: I1127 12:30:38.738445 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:30:38 crc kubenswrapper[4796]: I1127 12:30:38.739357 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-api" containerID="cri-o://5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb" gracePeriod=30 Nov 27 12:30:38 crc kubenswrapper[4796]: I1127 12:30:38.739443 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-notifier" containerID="cri-o://1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628" gracePeriod=30 Nov 27 12:30:38 crc kubenswrapper[4796]: I1127 12:30:38.739471 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-evaluator" containerID="cri-o://6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5" gracePeriod=30 Nov 27 12:30:38 crc kubenswrapper[4796]: I1127 12:30:38.739458 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-listener" containerID="cri-o://97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342" gracePeriod=30 Nov 27 12:30:39 crc kubenswrapper[4796]: I1127 12:30:39.521841 4796 generic.go:334] "Generic (PLEG): container finished" podID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerID="6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5" exitCode=0 Nov 27 12:30:39 crc kubenswrapper[4796]: I1127 12:30:39.522444 4796 generic.go:334] "Generic (PLEG): container finished" podID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerID="5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb" exitCode=0 Nov 27 12:30:39 crc kubenswrapper[4796]: I1127 12:30:39.521917 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerDied","Data":"6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5"} Nov 27 12:30:39 crc kubenswrapper[4796]: I1127 12:30:39.522558 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerDied","Data":"5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb"} Nov 27 12:30:39 crc kubenswrapper[4796]: I1127 12:30:39.569209 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:30:39 crc kubenswrapper[4796]: E1127 12:30:39.569715 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:30:42 crc kubenswrapper[4796]: I1127 12:30:42.551168 4796 generic.go:334] "Generic (PLEG): container finished" podID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerID="1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628" exitCode=0 Nov 27 12:30:42 crc kubenswrapper[4796]: I1127 12:30:42.551750 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerDied","Data":"1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628"} Nov 27 12:30:42 crc kubenswrapper[4796]: I1127 12:30:42.681776 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-t8f4v" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.075640 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.198611 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cng6p\" (UniqueName: \"kubernetes.io/projected/52f66cd2-21fb-4949-829f-1e05e6e010ee-kube-api-access-cng6p\") pod \"52f66cd2-21fb-4949-829f-1e05e6e010ee\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.198707 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-combined-ca-bundle\") pod \"52f66cd2-21fb-4949-829f-1e05e6e010ee\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.198788 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-scripts\") pod \"52f66cd2-21fb-4949-829f-1e05e6e010ee\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.198812 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-public-tls-certs\") pod \"52f66cd2-21fb-4949-829f-1e05e6e010ee\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.198879 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-internal-tls-certs\") pod \"52f66cd2-21fb-4949-829f-1e05e6e010ee\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.198934 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-config-data\") pod \"52f66cd2-21fb-4949-829f-1e05e6e010ee\" (UID: \"52f66cd2-21fb-4949-829f-1e05e6e010ee\") " Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.208892 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52f66cd2-21fb-4949-829f-1e05e6e010ee-kube-api-access-cng6p" (OuterVolumeSpecName: "kube-api-access-cng6p") pod "52f66cd2-21fb-4949-829f-1e05e6e010ee" (UID: "52f66cd2-21fb-4949-829f-1e05e6e010ee"). InnerVolumeSpecName "kube-api-access-cng6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.211193 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-scripts" (OuterVolumeSpecName: "scripts") pod "52f66cd2-21fb-4949-829f-1e05e6e010ee" (UID: "52f66cd2-21fb-4949-829f-1e05e6e010ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.262221 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "52f66cd2-21fb-4949-829f-1e05e6e010ee" (UID: "52f66cd2-21fb-4949-829f-1e05e6e010ee"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.276490 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "52f66cd2-21fb-4949-829f-1e05e6e010ee" (UID: "52f66cd2-21fb-4949-829f-1e05e6e010ee"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.301740 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.301806 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.301820 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.301832 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cng6p\" (UniqueName: \"kubernetes.io/projected/52f66cd2-21fb-4949-829f-1e05e6e010ee-kube-api-access-cng6p\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.312179 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-config-data" (OuterVolumeSpecName: "config-data") pod "52f66cd2-21fb-4949-829f-1e05e6e010ee" (UID: "52f66cd2-21fb-4949-829f-1e05e6e010ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.350612 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52f66cd2-21fb-4949-829f-1e05e6e010ee" (UID: "52f66cd2-21fb-4949-829f-1e05e6e010ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.403699 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.403776 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f66cd2-21fb-4949-829f-1e05e6e010ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.564163 4796 generic.go:334] "Generic (PLEG): container finished" podID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerID="97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342" exitCode=0 Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.564235 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.564249 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerDied","Data":"97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342"} Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.564333 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52f66cd2-21fb-4949-829f-1e05e6e010ee","Type":"ContainerDied","Data":"396b3e4e378dfe25701126bc0748cee8c8e621fdcd8981f111fdfae6a86d06ab"} Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.564395 4796 scope.go:117] "RemoveContainer" containerID="97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.566918 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" event={"ID":"12aebb1f-5956-4364-b2c9-80cf77ddb77a","Type":"ContainerStarted","Data":"91bb59bbc08d1425f0b0b1c4c11aa97b79c7780563bc5007c7d60c0d2431f627"} Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.587142 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-q7hl9" podStartSLOduration=4.190155373 podStartE2EDuration="32.587123691s" podCreationTimestamp="2025-11-27 12:30:11 +0000 UTC" firstStartedPulling="2025-11-27 12:30:13.214494726 +0000 UTC m=+3930.732813644" lastFinishedPulling="2025-11-27 12:30:41.611463034 +0000 UTC m=+3959.129781962" observedRunningTime="2025-11-27 12:30:43.582205277 +0000 UTC m=+3961.100524185" watchObservedRunningTime="2025-11-27 12:30:43.587123691 +0000 UTC m=+3961.105442609" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.605369 4796 scope.go:117] "RemoveContainer" containerID="1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.621323 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.638993 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.646117 4796 scope.go:117] "RemoveContainer" containerID="6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.647245 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.648016 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-evaluator" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648041 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-evaluator" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.648059 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-notifier" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648067 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-notifier" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.648097 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-api" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648105 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-api" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.648134 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-listener" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648143 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-listener" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648434 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-listener" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648460 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-evaluator" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648477 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-notifier" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.648499 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" containerName="aodh-api" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.650878 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.653819 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.656346 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vs2ld" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.656611 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.656778 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.659419 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.670150 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.670233 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice/crio-396b3e4e378dfe25701126bc0748cee8c8e621fdcd8981f111fdfae6a86d06ab\": RecentStats: unable to find data in memory cache]" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.683423 4796 scope.go:117] "RemoveContainer" containerID="5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.709162 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-scripts\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.709249 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.709345 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqwt6\" (UniqueName: \"kubernetes.io/projected/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-kube-api-access-qqwt6\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.709362 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-internal-tls-certs\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.709396 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-config-data\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.709448 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-public-tls-certs\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.713123 4796 scope.go:117] "RemoveContainer" containerID="97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.717165 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342\": container with ID starting with 97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342 not found: ID does not exist" containerID="97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.717231 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342"} err="failed to get container status \"97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342\": rpc error: code = NotFound desc = could not find container \"97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342\": container with ID starting with 97e313900956692a9d7931deec0a9a834acf9764094af4520ff55877e9357342 not found: ID does not exist" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.717278 4796 scope.go:117] "RemoveContainer" containerID="1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.717686 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628\": container with ID starting with 1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628 not found: ID does not exist" containerID="1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.717738 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628"} err="failed to get container status \"1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628\": rpc error: code = NotFound desc = could not find container \"1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628\": container with ID starting with 1e18d3545e904d839a0f01e43ab6bd4df57bf358019c3f04f2f5abf0f6e7c628 not found: ID does not exist" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.717766 4796 scope.go:117] "RemoveContainer" containerID="6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.720849 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5\": container with ID starting with 6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5 not found: ID does not exist" containerID="6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.720909 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5"} err="failed to get container status \"6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5\": rpc error: code = NotFound desc = could not find container \"6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5\": container with ID starting with 6155356730d183a2ff4d17251918b234e5d1741c9d98e584651b321f4c386cd5 not found: ID does not exist" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.720945 4796 scope.go:117] "RemoveContainer" containerID="5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb" Nov 27 12:30:43 crc kubenswrapper[4796]: E1127 12:30:43.722010 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb\": container with ID starting with 5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb not found: ID does not exist" containerID="5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.722050 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb"} err="failed to get container status \"5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb\": rpc error: code = NotFound desc = could not find container \"5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb\": container with ID starting with 5f2544db6a7e1233f2d3fe7d66e72826ca296643d095ed1bce20a7e5a0477efb not found: ID does not exist" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.811486 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-public-tls-certs\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.811599 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-scripts\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.811640 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.811683 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqwt6\" (UniqueName: \"kubernetes.io/projected/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-kube-api-access-qqwt6\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.811698 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-internal-tls-certs\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.811732 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-config-data\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.816699 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-scripts\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.817418 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.817614 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-config-data\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.817912 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-internal-tls-certs\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.817888 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-public-tls-certs\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.830317 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqwt6\" (UniqueName: \"kubernetes.io/projected/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-kube-api-access-qqwt6\") pod \"aodh-0\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " pod="openstack/aodh-0" Nov 27 12:30:43 crc kubenswrapper[4796]: I1127 12:30:43.974423 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:30:44 crc kubenswrapper[4796]: I1127 12:30:44.437580 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:30:44 crc kubenswrapper[4796]: W1127 12:30:44.441907 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d1e082f_a05f_41f5_93ff_613f85bd8f7e.slice/crio-a31102061ad4ab011caff33c7c2a700855fb0ce2597e64bc0c0d486ba753ceab WatchSource:0}: Error finding container a31102061ad4ab011caff33c7c2a700855fb0ce2597e64bc0c0d486ba753ceab: Status 404 returned error can't find the container with id a31102061ad4ab011caff33c7c2a700855fb0ce2597e64bc0c0d486ba753ceab Nov 27 12:30:44 crc kubenswrapper[4796]: I1127 12:30:44.579434 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerStarted","Data":"a31102061ad4ab011caff33c7c2a700855fb0ce2597e64bc0c0d486ba753ceab"} Nov 27 12:30:45 crc kubenswrapper[4796]: I1127 12:30:45.587429 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52f66cd2-21fb-4949-829f-1e05e6e010ee" path="/var/lib/kubelet/pods/52f66cd2-21fb-4949-829f-1e05e6e010ee/volumes" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.591137 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.593645 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.598589 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.598594 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.598686 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.598774 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-8z97v" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.600056 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.604252 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.619518 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerStarted","Data":"7b8323d76dc174141b2256f80e8d14050187672cd1debfc504d605f001b86028"} Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.619574 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerStarted","Data":"fcf03551c723527530c3fd439e2d1bfd1af5a4c858dc7c8aec5c13e947728d28"} Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666355 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666433 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4924a8c5-467a-47c0-9763-b10559f40519-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666564 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8zkn\" (UniqueName: \"kubernetes.io/projected/4924a8c5-467a-47c0-9763-b10559f40519-kube-api-access-z8zkn\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666590 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666663 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666685 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4924a8c5-467a-47c0-9763-b10559f40519-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.666711 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/4924a8c5-467a-47c0-9763-b10559f40519-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.768934 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8zkn\" (UniqueName: \"kubernetes.io/projected/4924a8c5-467a-47c0-9763-b10559f40519-kube-api-access-z8zkn\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.768999 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.769081 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.769109 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4924a8c5-467a-47c0-9763-b10559f40519-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.769152 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/4924a8c5-467a-47c0-9763-b10559f40519-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.769180 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.769228 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4924a8c5-467a-47c0-9763-b10559f40519-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.769809 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/4924a8c5-467a-47c0-9763-b10559f40519-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.773879 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.774060 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4924a8c5-467a-47c0-9763-b10559f40519-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.775472 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.776399 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4924a8c5-467a-47c0-9763-b10559f40519-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.787508 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8zkn\" (UniqueName: \"kubernetes.io/projected/4924a8c5-467a-47c0-9763-b10559f40519-kube-api-access-z8zkn\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.803322 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4924a8c5-467a-47c0-9763-b10559f40519-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"4924a8c5-467a-47c0-9763-b10559f40519\") " pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:46 crc kubenswrapper[4796]: I1127 12:30:46.921396 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 27 12:30:47 crc kubenswrapper[4796]: I1127 12:30:47.652527 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerStarted","Data":"1f0d186872d2585e6f778ce092698511e455391e975e51f39fa6a7751d7e5fcd"} Nov 27 12:30:47 crc kubenswrapper[4796]: I1127 12:30:47.766352 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 27 12:30:47 crc kubenswrapper[4796]: W1127 12:30:47.781188 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4924a8c5_467a_47c0_9763_b10559f40519.slice/crio-a47d0eb18dd56fdac81f5c485d23e8f7fe6e016ec30aae6d5ed161005875ca7d WatchSource:0}: Error finding container a47d0eb18dd56fdac81f5c485d23e8f7fe6e016ec30aae6d5ed161005875ca7d: Status 404 returned error can't find the container with id a47d0eb18dd56fdac81f5c485d23e8f7fe6e016ec30aae6d5ed161005875ca7d Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.050950 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.053483 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.055114 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.055208 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.056147 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.056259 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.056318 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6tv95" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.075148 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.076702 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208038 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x745f\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-kube-api-access-x745f\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208122 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208158 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208218 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208243 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208288 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208370 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.208446 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.309758 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.309879 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.309910 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x745f\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-kube-api-access-x745f\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.309957 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.309988 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.310047 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.310069 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.310108 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.311819 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.320326 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.321230 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.321396 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.322044 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.324589 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.324624 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4017d64cd4a9df3ce19ff345978cb75a46c1c617368a96f12e56b4cbcbada264/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.328135 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.337106 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x745f\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-kube-api-access-x745f\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.451258 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.671030 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.674996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"4924a8c5-467a-47c0-9763-b10559f40519","Type":"ContainerStarted","Data":"a47d0eb18dd56fdac81f5c485d23e8f7fe6e016ec30aae6d5ed161005875ca7d"} Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.677578 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerStarted","Data":"f28fe195189b0a349e742ab471cc8af5ec0bad3f2be5775660fcb9a285e5ab76"} Nov 27 12:30:48 crc kubenswrapper[4796]: I1127 12:30:48.706773 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.9395898150000002 podStartE2EDuration="5.706717498s" podCreationTimestamp="2025-11-27 12:30:43 +0000 UTC" firstStartedPulling="2025-11-27 12:30:44.444365482 +0000 UTC m=+3961.962684390" lastFinishedPulling="2025-11-27 12:30:48.211493155 +0000 UTC m=+3965.729812073" observedRunningTime="2025-11-27 12:30:48.694748041 +0000 UTC m=+3966.213066979" watchObservedRunningTime="2025-11-27 12:30:48.706717498 +0000 UTC m=+3966.225036426" Nov 27 12:30:49 crc kubenswrapper[4796]: I1127 12:30:49.256963 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:30:49 crc kubenswrapper[4796]: I1127 12:30:49.695452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerStarted","Data":"f2e394503199e4d12de9c47f8776598cd6de744b4960ebd95057f861c81b35b3"} Nov 27 12:30:51 crc kubenswrapper[4796]: I1127 12:30:51.569948 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:30:51 crc kubenswrapper[4796]: E1127 12:30:51.570605 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:30:53 crc kubenswrapper[4796]: I1127 12:30:53.795776 4796 scope.go:117] "RemoveContainer" containerID="d25589668379b9394e15cb36d64f57f7e46f13696ba3e7700b0682e04e827ca0" Nov 27 12:30:53 crc kubenswrapper[4796]: E1127 12:30:53.909545 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice\": RecentStats: unable to find data in memory cache]" Nov 27 12:30:54 crc kubenswrapper[4796]: I1127 12:30:54.757106 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"4924a8c5-467a-47c0-9763-b10559f40519","Type":"ContainerStarted","Data":"c838d8835e0d69c90e1423af19eac9b1ede9de0336f63a8ced39cf1eb0a7bfaf"} Nov 27 12:30:54 crc kubenswrapper[4796]: I1127 12:30:54.759101 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerStarted","Data":"52c34d67b2c0a8884f2e558bd777ee392938afd6ed94e188d15a3701d9667ba3"} Nov 27 12:31:01 crc kubenswrapper[4796]: I1127 12:31:01.834364 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerID="52c34d67b2c0a8884f2e558bd777ee392938afd6ed94e188d15a3701d9667ba3" exitCode=0 Nov 27 12:31:01 crc kubenswrapper[4796]: I1127 12:31:01.834492 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerDied","Data":"52c34d67b2c0a8884f2e558bd777ee392938afd6ed94e188d15a3701d9667ba3"} Nov 27 12:31:01 crc kubenswrapper[4796]: I1127 12:31:01.837110 4796 generic.go:334] "Generic (PLEG): container finished" podID="4924a8c5-467a-47c0-9763-b10559f40519" containerID="c838d8835e0d69c90e1423af19eac9b1ede9de0336f63a8ced39cf1eb0a7bfaf" exitCode=0 Nov 27 12:31:01 crc kubenswrapper[4796]: I1127 12:31:01.837174 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"4924a8c5-467a-47c0-9763-b10559f40519","Type":"ContainerDied","Data":"c838d8835e0d69c90e1423af19eac9b1ede9de0336f63a8ced39cf1eb0a7bfaf"} Nov 27 12:31:04 crc kubenswrapper[4796]: E1127 12:31:04.173454 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice\": RecentStats: unable to find data in memory cache]" Nov 27 12:31:04 crc kubenswrapper[4796]: I1127 12:31:04.867107 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"4924a8c5-467a-47c0-9763-b10559f40519","Type":"ContainerStarted","Data":"80ab8c2ed3381aa16f8de80401b2e998a5b9e1576b29aec52510d79bc6c8c1e0"} Nov 27 12:31:05 crc kubenswrapper[4796]: I1127 12:31:05.570708 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:31:05 crc kubenswrapper[4796]: E1127 12:31:05.571017 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:31:07 crc kubenswrapper[4796]: I1127 12:31:07.901212 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"4924a8c5-467a-47c0-9763-b10559f40519","Type":"ContainerStarted","Data":"a4f09a96a6ca786ce776bdf835b0b66b6931045276df220076ba584f3e735dad"} Nov 27 12:31:07 crc kubenswrapper[4796]: I1127 12:31:07.902108 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 27 12:31:07 crc kubenswrapper[4796]: I1127 12:31:07.905329 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 27 12:31:07 crc kubenswrapper[4796]: I1127 12:31:07.959572 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.878405391 podStartE2EDuration="21.959534363s" podCreationTimestamp="2025-11-27 12:30:46 +0000 UTC" firstStartedPulling="2025-11-27 12:30:47.798299844 +0000 UTC m=+3965.316618762" lastFinishedPulling="2025-11-27 12:31:03.879428816 +0000 UTC m=+3981.397747734" observedRunningTime="2025-11-27 12:31:07.928021866 +0000 UTC m=+3985.446340804" watchObservedRunningTime="2025-11-27 12:31:07.959534363 +0000 UTC m=+3985.477853421" Nov 27 12:31:08 crc kubenswrapper[4796]: I1127 12:31:08.915962 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerStarted","Data":"a3ba385f093bf246c649aa98ec835d925d8b4940c6b295f4bab46130c22f2085"} Nov 27 12:31:12 crc kubenswrapper[4796]: I1127 12:31:12.958557 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerStarted","Data":"9721658ec4577f84b88971aa26ac3261dcf45b39ff8abad40c3ee6a2198711d8"} Nov 27 12:31:14 crc kubenswrapper[4796]: E1127 12:31:14.434567 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice\": RecentStats: unable to find data in memory cache]" Nov 27 12:31:15 crc kubenswrapper[4796]: I1127 12:31:15.994408 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerStarted","Data":"b267aaeccc76699fa6c068f991e398951bb8556bc12a5bae83cfa6a2d11a875f"} Nov 27 12:31:16 crc kubenswrapper[4796]: I1127 12:31:16.042925 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.169252085 podStartE2EDuration="29.042901849s" podCreationTimestamp="2025-11-27 12:30:47 +0000 UTC" firstStartedPulling="2025-11-27 12:30:49.275066489 +0000 UTC m=+3966.793385407" lastFinishedPulling="2025-11-27 12:31:15.148716243 +0000 UTC m=+3992.667035171" observedRunningTime="2025-11-27 12:31:16.031630382 +0000 UTC m=+3993.549949320" watchObservedRunningTime="2025-11-27 12:31:16.042901849 +0000 UTC m=+3993.561220787" Nov 27 12:31:18 crc kubenswrapper[4796]: I1127 12:31:18.671511 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:18 crc kubenswrapper[4796]: I1127 12:31:18.672155 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:18 crc kubenswrapper[4796]: I1127 12:31:18.673783 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:19 crc kubenswrapper[4796]: I1127 12:31:19.030197 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.509498 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.510046 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" containerName="openstackclient" containerID="cri-o://27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350" gracePeriod=2 Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.528227 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.544837 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: E1127 12:31:20.545430 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" containerName="openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.545448 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" containerName="openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.545720 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" containerName="openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.546578 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.559247 4796 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T12:31:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T12:31:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T12:31:20Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T12:31:20Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\\\",\\\"name\\\":\\\"combined-ca-bundle\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bp8pf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T12:31:20Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.562175 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.569184 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:31:20 crc kubenswrapper[4796]: E1127 12:31:20.569458 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.583937 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: E1127 12:31:20.584763 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-bp8pf openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-bp8pf openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.622318 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.643709 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.645281 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.651015 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="527ec322-c155-4c49-ad66-74c5b82c595c" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.657633 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.666600 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.679527 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp8pf\" (UniqueName: \"kubernetes.io/projected/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-kube-api-access-bp8pf\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.679677 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.679751 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config-secret\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.679772 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.781442 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4bdf142a-761b-4879-b80b-b23c3f974593-openstack-config-secret\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.781760 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4bdf142a-761b-4879-b80b-b23c3f974593-openstack-config\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.781908 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.782060 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpgz2\" (UniqueName: \"kubernetes.io/projected/4bdf142a-761b-4879-b80b-b23c3f974593-kube-api-access-tpgz2\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.782112 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config-secret\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.782152 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.782198 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdf142a-761b-4879-b80b-b23c3f974593-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.782402 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp8pf\" (UniqueName: \"kubernetes.io/projected/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-kube-api-access-bp8pf\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.783534 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: E1127 12:31:20.786496 4796 projected.go:194] Error preparing data for projected volume kube-api-access-bp8pf for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (5ac24dcf-b27a-4542-a697-e9fada9c2cfb) does not match the UID in record. The object might have been deleted and then recreated Nov 27 12:31:20 crc kubenswrapper[4796]: E1127 12:31:20.786590 4796 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-kube-api-access-bp8pf podName:5ac24dcf-b27a-4542-a697-e9fada9c2cfb nodeName:}" failed. No retries permitted until 2025-11-27 12:31:21.286570248 +0000 UTC m=+3998.804889156 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bp8pf" (UniqueName: "kubernetes.io/projected/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-kube-api-access-bp8pf") pod "openstackclient" (UID: "5ac24dcf-b27a-4542-a697-e9fada9c2cfb") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (5ac24dcf-b27a-4542-a697-e9fada9c2cfb) does not match the UID in record. The object might have been deleted and then recreated Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.883691 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4bdf142a-761b-4879-b80b-b23c3f974593-openstack-config\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.883799 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpgz2\" (UniqueName: \"kubernetes.io/projected/4bdf142a-761b-4879-b80b-b23c3f974593-kube-api-access-tpgz2\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.883840 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdf142a-761b-4879-b80b-b23c3f974593-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.883918 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4bdf142a-761b-4879-b80b-b23c3f974593-openstack-config-secret\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.885103 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4bdf142a-761b-4879-b80b-b23c3f974593-openstack-config\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.891772 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.892126 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-api" containerID="cri-o://fcf03551c723527530c3fd439e2d1bfd1af5a4c858dc7c8aec5c13e947728d28" gracePeriod=30 Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.892533 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-evaluator" containerID="cri-o://7b8323d76dc174141b2256f80e8d14050187672cd1debfc504d605f001b86028" gracePeriod=30 Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.892562 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-notifier" containerID="cri-o://1f0d186872d2585e6f778ce092698511e455391e975e51f39fa6a7751d7e5fcd" gracePeriod=30 Nov 27 12:31:20 crc kubenswrapper[4796]: I1127 12:31:20.892707 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-listener" containerID="cri-o://f28fe195189b0a349e742ab471cc8af5ec0bad3f2be5775660fcb9a285e5ab76" gracePeriod=30 Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.062616 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.073168 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.082064 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.084772 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.189398 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config\") pod \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.189914 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5ac24dcf-b27a-4542-a697-e9fada9c2cfb" (UID: "5ac24dcf-b27a-4542-a697-e9fada9c2cfb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.190806 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.190834 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp8pf\" (UniqueName: \"kubernetes.io/projected/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-kube-api-access-bp8pf\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.269287 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config-secret\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.269790 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.269974 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4bdf142a-761b-4879-b80b-b23c3f974593-openstack-config-secret\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.270575 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpgz2\" (UniqueName: \"kubernetes.io/projected/4bdf142a-761b-4879-b80b-b23c3f974593-kube-api-access-tpgz2\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.270620 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdf142a-761b-4879-b80b-b23c3f974593-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4bdf142a-761b-4879-b80b-b23c3f974593\") " pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.286045 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.291902 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-combined-ca-bundle\") pod \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.292128 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config-secret\") pod \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\" (UID: \"5ac24dcf-b27a-4542-a697-e9fada9c2cfb\") " Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.371495 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5ac24dcf-b27a-4542-a697-e9fada9c2cfb" (UID: "5ac24dcf-b27a-4542-a697-e9fada9c2cfb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.377437 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ac24dcf-b27a-4542-a697-e9fada9c2cfb" (UID: "5ac24dcf-b27a-4542-a697-e9fada9c2cfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.399986 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.400243 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5ac24dcf-b27a-4542-a697-e9fada9c2cfb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.593097 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" path="/var/lib/kubelet/pods/5ac24dcf-b27a-4542-a697-e9fada9c2cfb/volumes" Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.708904 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.709385 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="prometheus" containerID="cri-o://a3ba385f093bf246c649aa98ec835d925d8b4940c6b295f4bab46130c22f2085" gracePeriod=600 Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.709932 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="thanos-sidecar" containerID="cri-o://b267aaeccc76699fa6c068f991e398951bb8556bc12a5bae83cfa6a2d11a875f" gracePeriod=600 Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.709999 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="config-reloader" containerID="cri-o://9721658ec4577f84b88971aa26ac3261dcf45b39ff8abad40c3ee6a2198711d8" gracePeriod=600 Nov 27 12:31:21 crc kubenswrapper[4796]: W1127 12:31:21.806815 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bdf142a_761b_4879_b80b_b23c3f974593.slice/crio-86c6bfaff81d1650613d872b76ad7024f621eead885c4c38b93a0a8ed0003b80 WatchSource:0}: Error finding container 86c6bfaff81d1650613d872b76ad7024f621eead885c4c38b93a0a8ed0003b80: Status 404 returned error can't find the container with id 86c6bfaff81d1650613d872b76ad7024f621eead885c4c38b93a0a8ed0003b80 Nov 27 12:31:21 crc kubenswrapper[4796]: I1127 12:31:21.808217 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.075373 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerID="b267aaeccc76699fa6c068f991e398951bb8556bc12a5bae83cfa6a2d11a875f" exitCode=0 Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.075407 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerID="9721658ec4577f84b88971aa26ac3261dcf45b39ff8abad40c3ee6a2198711d8" exitCode=0 Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.075419 4796 generic.go:334] "Generic (PLEG): container finished" podID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerID="a3ba385f093bf246c649aa98ec835d925d8b4940c6b295f4bab46130c22f2085" exitCode=0 Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.075423 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerDied","Data":"b267aaeccc76699fa6c068f991e398951bb8556bc12a5bae83cfa6a2d11a875f"} Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.075488 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerDied","Data":"9721658ec4577f84b88971aa26ac3261dcf45b39ff8abad40c3ee6a2198711d8"} Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.075503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerDied","Data":"a3ba385f093bf246c649aa98ec835d925d8b4940c6b295f4bab46130c22f2085"} Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.078344 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4bdf142a-761b-4879-b80b-b23c3f974593","Type":"ContainerStarted","Data":"86c6bfaff81d1650613d872b76ad7024f621eead885c4c38b93a0a8ed0003b80"} Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.083417 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerID="7b8323d76dc174141b2256f80e8d14050187672cd1debfc504d605f001b86028" exitCode=0 Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.083454 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerID="fcf03551c723527530c3fd439e2d1bfd1af5a4c858dc7c8aec5c13e947728d28" exitCode=0 Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.083503 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.083498 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerDied","Data":"7b8323d76dc174141b2256f80e8d14050187672cd1debfc504d605f001b86028"} Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.083538 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerDied","Data":"fcf03551c723527530c3fd439e2d1bfd1af5a4c858dc7c8aec5c13e947728d28"} Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.085980 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.089231 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="5ac24dcf-b27a-4542-a697-e9fada9c2cfb" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.648585 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.728933 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-tls-assets\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729020 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-web-config\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729095 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-prometheus-metric-storage-rulefiles-0\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729318 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729392 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x745f\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-kube-api-access-x745f\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729441 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-thanos-prometheus-http-client-file\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729509 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729602 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config-out\") pod \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\" (UID: \"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1\") " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.729928 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.730441 4796 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.737482 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-kube-api-access-x745f" (OuterVolumeSpecName: "kube-api-access-x745f") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "kube-api-access-x745f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.745322 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.749923 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config-out" (OuterVolumeSpecName: "config-out") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.758867 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config" (OuterVolumeSpecName: "config") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.765794 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.777894 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-web-config" (OuterVolumeSpecName: "web-config") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.789778 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" (UID: "eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1"). InnerVolumeSpecName "pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833666 4796 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833710 4796 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-web-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833761 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") on node \"crc\" " Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833779 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x745f\" (UniqueName: \"kubernetes.io/projected/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-kube-api-access-x745f\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833795 4796 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833808 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.833819 4796 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1-config-out\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.868095 4796 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.868514 4796 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9") on node "crc" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.875937 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:22 crc kubenswrapper[4796]: I1127 12:31:22.936571 4796 reconciler_common.go:293] "Volume detached for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.037639 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-combined-ca-bundle\") pod \"527ec322-c155-4c49-ad66-74c5b82c595c\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.037694 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config-secret\") pod \"527ec322-c155-4c49-ad66-74c5b82c595c\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.037771 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config\") pod \"527ec322-c155-4c49-ad66-74c5b82c595c\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.037810 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngg4g\" (UniqueName: \"kubernetes.io/projected/527ec322-c155-4c49-ad66-74c5b82c595c-kube-api-access-ngg4g\") pod \"527ec322-c155-4c49-ad66-74c5b82c595c\" (UID: \"527ec322-c155-4c49-ad66-74c5b82c595c\") " Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.042112 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527ec322-c155-4c49-ad66-74c5b82c595c-kube-api-access-ngg4g" (OuterVolumeSpecName: "kube-api-access-ngg4g") pod "527ec322-c155-4c49-ad66-74c5b82c595c" (UID: "527ec322-c155-4c49-ad66-74c5b82c595c"). InnerVolumeSpecName "kube-api-access-ngg4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.074593 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "527ec322-c155-4c49-ad66-74c5b82c595c" (UID: "527ec322-c155-4c49-ad66-74c5b82c595c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.084488 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "527ec322-c155-4c49-ad66-74c5b82c595c" (UID: "527ec322-c155-4c49-ad66-74c5b82c595c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.095836 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "527ec322-c155-4c49-ad66-74c5b82c595c" (UID: "527ec322-c155-4c49-ad66-74c5b82c595c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.098757 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1","Type":"ContainerDied","Data":"f2e394503199e4d12de9c47f8776598cd6de744b4960ebd95057f861c81b35b3"} Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.098812 4796 scope.go:117] "RemoveContainer" containerID="b267aaeccc76699fa6c068f991e398951bb8556bc12a5bae83cfa6a2d11a875f" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.098770 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.101716 4796 generic.go:334] "Generic (PLEG): container finished" podID="527ec322-c155-4c49-ad66-74c5b82c595c" containerID="27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350" exitCode=137 Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.101843 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.107381 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4bdf142a-761b-4879-b80b-b23c3f974593","Type":"ContainerStarted","Data":"1884b628e894fcdf19d2f58edf9c0f0368a2df9f589b5df451f4ae8cde4ed24b"} Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.140230 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.140258 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.140282 4796 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/527ec322-c155-4c49-ad66-74c5b82c595c-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.140293 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngg4g\" (UniqueName: \"kubernetes.io/projected/527ec322-c155-4c49-ad66-74c5b82c595c-kube-api-access-ngg4g\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.160350 4796 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="527ec322-c155-4c49-ad66-74c5b82c595c" podUID="4bdf142a-761b-4879-b80b-b23c3f974593" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.161574 4796 scope.go:117] "RemoveContainer" containerID="9721658ec4577f84b88971aa26ac3261dcf45b39ff8abad40c3ee6a2198711d8" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.183068 4796 scope.go:117] "RemoveContainer" containerID="a3ba385f093bf246c649aa98ec835d925d8b4940c6b295f4bab46130c22f2085" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.186973 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.186956699 podStartE2EDuration="3.186956699s" podCreationTimestamp="2025-11-27 12:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:31:23.127016789 +0000 UTC m=+4000.645335707" watchObservedRunningTime="2025-11-27 12:31:23.186956699 +0000 UTC m=+4000.705275617" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.187899 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.201227 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.212278 4796 scope.go:117] "RemoveContainer" containerID="52c34d67b2c0a8884f2e558bd777ee392938afd6ed94e188d15a3701d9667ba3" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.215943 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:31:23 crc kubenswrapper[4796]: E1127 12:31:23.216351 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="prometheus" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216369 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="prometheus" Nov 27 12:31:23 crc kubenswrapper[4796]: E1127 12:31:23.216384 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="thanos-sidecar" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216391 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="thanos-sidecar" Nov 27 12:31:23 crc kubenswrapper[4796]: E1127 12:31:23.216409 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="config-reloader" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216421 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="config-reloader" Nov 27 12:31:23 crc kubenswrapper[4796]: E1127 12:31:23.216448 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="init-config-reloader" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216454 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="init-config-reloader" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216666 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="prometheus" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216696 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="config-reloader" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.216708 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" containerName="thanos-sidecar" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.218774 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.224507 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.224708 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.224760 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.224895 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.225007 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6tv95" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.225141 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.231506 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.237328 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.238418 4796 scope.go:117] "RemoveContainer" containerID="27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.279642 4796 scope.go:117] "RemoveContainer" containerID="27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350" Nov 27 12:31:23 crc kubenswrapper[4796]: E1127 12:31:23.281092 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350\": container with ID starting with 27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350 not found: ID does not exist" containerID="27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.281137 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350"} err="failed to get container status \"27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350\": rpc error: code = NotFound desc = could not find container \"27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350\": container with ID starting with 27530a6a47842991c6cc3cff2d347c8de332f2e04442f59f83f63e71d5fcd350 not found: ID does not exist" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.345821 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.345919 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.345950 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346082 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0383092-f637-4b82-a767-bf53ea59209e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346107 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346242 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-config\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346372 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a0383092-f637-4b82-a767-bf53ea59209e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346488 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vchl\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-kube-api-access-9vchl\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346591 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346775 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.346839 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450082 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450201 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450241 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450355 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0383092-f637-4b82-a767-bf53ea59209e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450390 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450426 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-config\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450475 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a0383092-f637-4b82-a767-bf53ea59209e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450536 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vchl\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-kube-api-access-9vchl\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450645 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450763 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.450832 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.451488 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a0383092-f637-4b82-a767-bf53ea59209e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.455788 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.455964 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-config\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.456033 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.456455 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0383092-f637-4b82-a767-bf53ea59209e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.458389 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.459053 4796 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.459160 4796 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4017d64cd4a9df3ce19ff345978cb75a46c1c617368a96f12e56b4cbcbada264/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.459651 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.460545 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.460850 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.469937 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vchl\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-kube-api-access-9vchl\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.507942 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"prometheus-metric-storage-0\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.544380 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.600511 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527ec322-c155-4c49-ad66-74c5b82c595c" path="/var/lib/kubelet/pods/527ec322-c155-4c49-ad66-74c5b82c595c/volumes" Nov 27 12:31:23 crc kubenswrapper[4796]: I1127 12:31:23.601976 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1" path="/var/lib/kubelet/pods/eb22f8ff-b6f9-41b5-bf30-1ab343b2c5f1/volumes" Nov 27 12:31:24 crc kubenswrapper[4796]: I1127 12:31:24.008520 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:31:24 crc kubenswrapper[4796]: I1127 12:31:24.118096 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerStarted","Data":"a5c352651f14edbf93ea78ea2cddc21cd33338540831229f48bc6a246e146772"} Nov 27 12:31:24 crc kubenswrapper[4796]: E1127 12:31:24.711688 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice\": RecentStats: unable to find data in memory cache]" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.250736 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerID="f28fe195189b0a349e742ab471cc8af5ec0bad3f2be5775660fcb9a285e5ab76" exitCode=0 Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.251315 4796 generic.go:334] "Generic (PLEG): container finished" podID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerID="1f0d186872d2585e6f778ce092698511e455391e975e51f39fa6a7751d7e5fcd" exitCode=0 Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.251340 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerDied","Data":"f28fe195189b0a349e742ab471cc8af5ec0bad3f2be5775660fcb9a285e5ab76"} Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.251380 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerDied","Data":"1f0d186872d2585e6f778ce092698511e455391e975e51f39fa6a7751d7e5fcd"} Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.470104 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.621764 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-public-tls-certs\") pod \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.622146 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-combined-ca-bundle\") pod \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.622168 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-internal-tls-certs\") pod \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.622200 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-config-data\") pod \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.622295 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-scripts\") pod \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.622386 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqwt6\" (UniqueName: \"kubernetes.io/projected/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-kube-api-access-qqwt6\") pod \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\" (UID: \"8d1e082f-a05f-41f5-93ff-613f85bd8f7e\") " Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.628490 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-scripts" (OuterVolumeSpecName: "scripts") pod "8d1e082f-a05f-41f5-93ff-613f85bd8f7e" (UID: "8d1e082f-a05f-41f5-93ff-613f85bd8f7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.628865 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-kube-api-access-qqwt6" (OuterVolumeSpecName: "kube-api-access-qqwt6") pod "8d1e082f-a05f-41f5-93ff-613f85bd8f7e" (UID: "8d1e082f-a05f-41f5-93ff-613f85bd8f7e"). InnerVolumeSpecName "kube-api-access-qqwt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.679999 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d1e082f-a05f-41f5-93ff-613f85bd8f7e" (UID: "8d1e082f-a05f-41f5-93ff-613f85bd8f7e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.693980 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8d1e082f-a05f-41f5-93ff-613f85bd8f7e" (UID: "8d1e082f-a05f-41f5-93ff-613f85bd8f7e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.725360 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.725387 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.725397 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.725406 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqwt6\" (UniqueName: \"kubernetes.io/projected/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-kube-api-access-qqwt6\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.731101 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d1e082f-a05f-41f5-93ff-613f85bd8f7e" (UID: "8d1e082f-a05f-41f5-93ff-613f85bd8f7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.756279 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-config-data" (OuterVolumeSpecName: "config-data") pod "8d1e082f-a05f-41f5-93ff-613f85bd8f7e" (UID: "8d1e082f-a05f-41f5-93ff-613f85bd8f7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.827033 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:26 crc kubenswrapper[4796]: I1127 12:31:26.827096 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d1e082f-a05f-41f5-93ff-613f85bd8f7e-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.264381 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8d1e082f-a05f-41f5-93ff-613f85bd8f7e","Type":"ContainerDied","Data":"a31102061ad4ab011caff33c7c2a700855fb0ce2597e64bc0c0d486ba753ceab"} Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.264444 4796 scope.go:117] "RemoveContainer" containerID="f28fe195189b0a349e742ab471cc8af5ec0bad3f2be5775660fcb9a285e5ab76" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.264539 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.295667 4796 scope.go:117] "RemoveContainer" containerID="1f0d186872d2585e6f778ce092698511e455391e975e51f39fa6a7751d7e5fcd" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.342881 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.350421 4796 scope.go:117] "RemoveContainer" containerID="7b8323d76dc174141b2256f80e8d14050187672cd1debfc504d605f001b86028" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.368477 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.372985 4796 scope.go:117] "RemoveContainer" containerID="fcf03551c723527530c3fd439e2d1bfd1af5a4c858dc7c8aec5c13e947728d28" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.379992 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 27 12:31:27 crc kubenswrapper[4796]: E1127 12:31:27.380766 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-evaluator" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.380793 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-evaluator" Nov 27 12:31:27 crc kubenswrapper[4796]: E1127 12:31:27.380804 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-notifier" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.380810 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-notifier" Nov 27 12:31:27 crc kubenswrapper[4796]: E1127 12:31:27.380829 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-listener" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.380836 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-listener" Nov 27 12:31:27 crc kubenswrapper[4796]: E1127 12:31:27.380848 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-api" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.380854 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-api" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.381055 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-listener" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.381071 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-evaluator" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.381089 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-notifier" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.381100 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" containerName="aodh-api" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.385077 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.390296 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.390331 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.390548 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.390667 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vs2ld" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.390750 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.392437 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.546417 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.546830 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-config-data\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.546863 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-scripts\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.546912 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/220303ef-0c3c-4c74-959a-129565095e51-kube-api-access-vchhk\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.546958 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-internal-tls-certs\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.546991 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-combined-ca-bundle\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.579001 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1e082f-a05f-41f5-93ff-613f85bd8f7e" path="/var/lib/kubelet/pods/8d1e082f-a05f-41f5-93ff-613f85bd8f7e/volumes" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.648714 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.648792 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-config-data\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.648819 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-scripts\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.648862 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/220303ef-0c3c-4c74-959a-129565095e51-kube-api-access-vchhk\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.648899 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-internal-tls-certs\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.648942 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-combined-ca-bundle\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.652718 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-combined-ca-bundle\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.652870 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.653014 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-scripts\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.653189 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-config-data\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.653866 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-internal-tls-certs\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.665607 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/220303ef-0c3c-4c74-959a-129565095e51-kube-api-access-vchhk\") pod \"aodh-0\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " pod="openstack/aodh-0" Nov 27 12:31:27 crc kubenswrapper[4796]: I1127 12:31:27.709620 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:31:28 crc kubenswrapper[4796]: I1127 12:31:28.165786 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:31:28 crc kubenswrapper[4796]: I1127 12:31:28.276256 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerStarted","Data":"ec75cc995c6c01f9624ea93cd210295df643cd96af2d8d02e8a675b67d4002d3"} Nov 27 12:31:28 crc kubenswrapper[4796]: I1127 12:31:28.278138 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerStarted","Data":"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3"} Nov 27 12:31:30 crc kubenswrapper[4796]: I1127 12:31:30.307387 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerStarted","Data":"39aaaf3c94179a85c88b3e5bba465336efc123996659a405bfaae7b9f2278475"} Nov 27 12:31:31 crc kubenswrapper[4796]: I1127 12:31:31.319477 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerStarted","Data":"5d576b837960794610df67aa6caa98c973f88e660b4def1ab1732453d33a5b2f"} Nov 27 12:31:32 crc kubenswrapper[4796]: I1127 12:31:32.332911 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerStarted","Data":"a4e657b598b67d12de4572a98206413e70744b00b8583ed9b8dab3a970415cfa"} Nov 27 12:31:32 crc kubenswrapper[4796]: I1127 12:31:32.569959 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:31:32 crc kubenswrapper[4796]: E1127 12:31:32.570207 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:31:33 crc kubenswrapper[4796]: I1127 12:31:33.344320 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerStarted","Data":"b9fe67225938676744c90f71d643031ffec9ceafd68ca980dff4391ec99f9a76"} Nov 27 12:31:33 crc kubenswrapper[4796]: I1127 12:31:33.374221 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.140601955 podStartE2EDuration="6.374185899s" podCreationTimestamp="2025-11-27 12:31:27 +0000 UTC" firstStartedPulling="2025-11-27 12:31:28.168197522 +0000 UTC m=+4005.686516450" lastFinishedPulling="2025-11-27 12:31:32.401781476 +0000 UTC m=+4009.920100394" observedRunningTime="2025-11-27 12:31:33.35986887 +0000 UTC m=+4010.878187828" watchObservedRunningTime="2025-11-27 12:31:33.374185899 +0000 UTC m=+4010.892504807" Nov 27 12:31:34 crc kubenswrapper[4796]: I1127 12:31:34.355345 4796 generic.go:334] "Generic (PLEG): container finished" podID="a0383092-f637-4b82-a767-bf53ea59209e" containerID="86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3" exitCode=0 Nov 27 12:31:34 crc kubenswrapper[4796]: I1127 12:31:34.356884 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerDied","Data":"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3"} Nov 27 12:31:34 crc kubenswrapper[4796]: E1127 12:31:34.970440 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52f66cd2_21fb_4949_829f_1e05e6e010ee.slice\": RecentStats: unable to find data in memory cache]" Nov 27 12:31:35 crc kubenswrapper[4796]: I1127 12:31:35.367568 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerStarted","Data":"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba"} Nov 27 12:31:38 crc kubenswrapper[4796]: I1127 12:31:38.406306 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerStarted","Data":"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404"} Nov 27 12:31:39 crc kubenswrapper[4796]: I1127 12:31:39.423320 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerStarted","Data":"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1"} Nov 27 12:31:39 crc kubenswrapper[4796]: I1127 12:31:39.471295 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.471238018 podStartE2EDuration="16.471238018s" podCreationTimestamp="2025-11-27 12:31:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:31:39.468874003 +0000 UTC m=+4016.987192961" watchObservedRunningTime="2025-11-27 12:31:39.471238018 +0000 UTC m=+4016.989556966" Nov 27 12:31:43 crc kubenswrapper[4796]: I1127 12:31:43.545144 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:43 crc kubenswrapper[4796]: I1127 12:31:43.579621 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:31:43 crc kubenswrapper[4796]: E1127 12:31:43.580015 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:31:53 crc kubenswrapper[4796]: I1127 12:31:53.545353 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:53 crc kubenswrapper[4796]: I1127 12:31:53.552615 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:53 crc kubenswrapper[4796]: I1127 12:31:53.606912 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 27 12:31:57 crc kubenswrapper[4796]: I1127 12:31:57.570042 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:31:57 crc kubenswrapper[4796]: E1127 12:31:57.570963 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:32:10 crc kubenswrapper[4796]: I1127 12:32:10.568828 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:32:10 crc kubenswrapper[4796]: E1127 12:32:10.569593 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:32:24 crc kubenswrapper[4796]: I1127 12:32:24.570902 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:32:24 crc kubenswrapper[4796]: E1127 12:32:24.572336 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:32:36 crc kubenswrapper[4796]: I1127 12:32:36.568621 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:32:36 crc kubenswrapper[4796]: E1127 12:32:36.569371 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.683910 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-skkp9"] Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.687817 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.702739 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-skkp9"] Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.761883 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-utilities\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.762224 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-catalog-content\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.762444 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9czv\" (UniqueName: \"kubernetes.io/projected/a4ebbefc-c2f7-4c6d-9442-958d387621f0-kube-api-access-b9czv\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.863765 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-catalog-content\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.863871 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9czv\" (UniqueName: \"kubernetes.io/projected/a4ebbefc-c2f7-4c6d-9442-958d387621f0-kube-api-access-b9czv\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.863980 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-utilities\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.864370 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-catalog-content\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.864779 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-utilities\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:41 crc kubenswrapper[4796]: I1127 12:32:41.891847 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9czv\" (UniqueName: \"kubernetes.io/projected/a4ebbefc-c2f7-4c6d-9442-958d387621f0-kube-api-access-b9czv\") pod \"community-operators-skkp9\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:42 crc kubenswrapper[4796]: I1127 12:32:42.012660 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:42 crc kubenswrapper[4796]: I1127 12:32:42.537354 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-skkp9"] Nov 27 12:32:43 crc kubenswrapper[4796]: I1127 12:32:43.010985 4796 generic.go:334] "Generic (PLEG): container finished" podID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerID="0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb" exitCode=0 Nov 27 12:32:43 crc kubenswrapper[4796]: I1127 12:32:43.011289 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerDied","Data":"0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb"} Nov 27 12:32:43 crc kubenswrapper[4796]: I1127 12:32:43.011319 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerStarted","Data":"7cbd18cf4725f18a125ba1c32fa18076d439914b90b2154d4ffb82c68defb449"} Nov 27 12:32:45 crc kubenswrapper[4796]: I1127 12:32:45.036433 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerStarted","Data":"a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f"} Nov 27 12:32:47 crc kubenswrapper[4796]: I1127 12:32:47.067581 4796 generic.go:334] "Generic (PLEG): container finished" podID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerID="a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f" exitCode=0 Nov 27 12:32:47 crc kubenswrapper[4796]: I1127 12:32:47.068263 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerDied","Data":"a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f"} Nov 27 12:32:47 crc kubenswrapper[4796]: I1127 12:32:47.569172 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:32:47 crc kubenswrapper[4796]: E1127 12:32:47.569546 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:32:48 crc kubenswrapper[4796]: I1127 12:32:48.080673 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerStarted","Data":"2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9"} Nov 27 12:32:48 crc kubenswrapper[4796]: I1127 12:32:48.108376 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-skkp9" podStartSLOduration=2.309983079 podStartE2EDuration="7.108353701s" podCreationTimestamp="2025-11-27 12:32:41 +0000 UTC" firstStartedPulling="2025-11-27 12:32:43.013702599 +0000 UTC m=+4080.532021517" lastFinishedPulling="2025-11-27 12:32:47.812073211 +0000 UTC m=+4085.330392139" observedRunningTime="2025-11-27 12:32:48.097296451 +0000 UTC m=+4085.615615369" watchObservedRunningTime="2025-11-27 12:32:48.108353701 +0000 UTC m=+4085.626672629" Nov 27 12:32:52 crc kubenswrapper[4796]: I1127 12:32:52.013106 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:52 crc kubenswrapper[4796]: I1127 12:32:52.013498 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:52 crc kubenswrapper[4796]: I1127 12:32:52.064494 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:32:58 crc kubenswrapper[4796]: I1127 12:32:58.569349 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:32:58 crc kubenswrapper[4796]: E1127 12:32:58.570010 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:33:02 crc kubenswrapper[4796]: I1127 12:33:02.225315 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:33:02 crc kubenswrapper[4796]: I1127 12:33:02.282666 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-skkp9"] Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.239293 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-skkp9" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="registry-server" containerID="cri-o://2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9" gracePeriod=2 Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.822833 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.922148 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-catalog-content\") pod \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.922224 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9czv\" (UniqueName: \"kubernetes.io/projected/a4ebbefc-c2f7-4c6d-9442-958d387621f0-kube-api-access-b9czv\") pod \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.922615 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-utilities\") pod \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\" (UID: \"a4ebbefc-c2f7-4c6d-9442-958d387621f0\") " Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.923402 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-utilities" (OuterVolumeSpecName: "utilities") pod "a4ebbefc-c2f7-4c6d-9442-958d387621f0" (UID: "a4ebbefc-c2f7-4c6d-9442-958d387621f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.927806 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ebbefc-c2f7-4c6d-9442-958d387621f0-kube-api-access-b9czv" (OuterVolumeSpecName: "kube-api-access-b9czv") pod "a4ebbefc-c2f7-4c6d-9442-958d387621f0" (UID: "a4ebbefc-c2f7-4c6d-9442-958d387621f0"). InnerVolumeSpecName "kube-api-access-b9czv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:33:03 crc kubenswrapper[4796]: I1127 12:33:03.972424 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4ebbefc-c2f7-4c6d-9442-958d387621f0" (UID: "a4ebbefc-c2f7-4c6d-9442-958d387621f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.025511 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.025557 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ebbefc-c2f7-4c6d-9442-958d387621f0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.025574 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9czv\" (UniqueName: \"kubernetes.io/projected/a4ebbefc-c2f7-4c6d-9442-958d387621f0-kube-api-access-b9czv\") on node \"crc\" DevicePath \"\"" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.251754 4796 generic.go:334] "Generic (PLEG): container finished" podID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerID="2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9" exitCode=0 Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.251851 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skkp9" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.251859 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerDied","Data":"2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9"} Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.252177 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skkp9" event={"ID":"a4ebbefc-c2f7-4c6d-9442-958d387621f0","Type":"ContainerDied","Data":"7cbd18cf4725f18a125ba1c32fa18076d439914b90b2154d4ffb82c68defb449"} Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.252224 4796 scope.go:117] "RemoveContainer" containerID="2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.280566 4796 scope.go:117] "RemoveContainer" containerID="a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.286594 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-skkp9"] Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.298590 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-skkp9"] Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.306512 4796 scope.go:117] "RemoveContainer" containerID="0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.371848 4796 scope.go:117] "RemoveContainer" containerID="2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9" Nov 27 12:33:04 crc kubenswrapper[4796]: E1127 12:33:04.372353 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9\": container with ID starting with 2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9 not found: ID does not exist" containerID="2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.372400 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9"} err="failed to get container status \"2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9\": rpc error: code = NotFound desc = could not find container \"2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9\": container with ID starting with 2170eee31971ef4031a780a736d4f281e50d727935e4372af062b2175d8cf7c9 not found: ID does not exist" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.372428 4796 scope.go:117] "RemoveContainer" containerID="a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f" Nov 27 12:33:04 crc kubenswrapper[4796]: E1127 12:33:04.373205 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f\": container with ID starting with a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f not found: ID does not exist" containerID="a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.373247 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f"} err="failed to get container status \"a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f\": rpc error: code = NotFound desc = could not find container \"a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f\": container with ID starting with a6258fa7379b7f8e4f025f7263cbb3022e8951eca23c975bcc446d7c2729817f not found: ID does not exist" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.373288 4796 scope.go:117] "RemoveContainer" containerID="0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb" Nov 27 12:33:04 crc kubenswrapper[4796]: E1127 12:33:04.374554 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb\": container with ID starting with 0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb not found: ID does not exist" containerID="0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb" Nov 27 12:33:04 crc kubenswrapper[4796]: I1127 12:33:04.374581 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb"} err="failed to get container status \"0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb\": rpc error: code = NotFound desc = could not find container \"0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb\": container with ID starting with 0339958cac3f9d4faf491bcfe596ae25b8dad46dfcf12aa90486b050761448bb not found: ID does not exist" Nov 27 12:33:05 crc kubenswrapper[4796]: I1127 12:33:05.597874 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" path="/var/lib/kubelet/pods/a4ebbefc-c2f7-4c6d-9442-958d387621f0/volumes" Nov 27 12:33:10 crc kubenswrapper[4796]: I1127 12:33:10.569162 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:33:10 crc kubenswrapper[4796]: E1127 12:33:10.570126 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:33:23 crc kubenswrapper[4796]: I1127 12:33:23.580583 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:33:23 crc kubenswrapper[4796]: E1127 12:33:23.581396 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:33:37 crc kubenswrapper[4796]: I1127 12:33:37.569004 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:33:37 crc kubenswrapper[4796]: E1127 12:33:37.569850 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:33:50 crc kubenswrapper[4796]: I1127 12:33:50.569902 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:33:50 crc kubenswrapper[4796]: E1127 12:33:50.570811 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:33:58 crc kubenswrapper[4796]: I1127 12:33:58.152581 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:33:59 crc kubenswrapper[4796]: I1127 12:33:59.845552 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:33:59 crc kubenswrapper[4796]: I1127 12:33:59.846056 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="prometheus" containerID="cri-o://4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" gracePeriod=600 Nov 27 12:33:59 crc kubenswrapper[4796]: I1127 12:33:59.846461 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="thanos-sidecar" containerID="cri-o://f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" gracePeriod=600 Nov 27 12:33:59 crc kubenswrapper[4796]: I1127 12:33:59.846519 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="config-reloader" containerID="cri-o://371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" gracePeriod=600 Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.808055 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831214 4796 generic.go:334] "Generic (PLEG): container finished" podID="a0383092-f637-4b82-a767-bf53ea59209e" containerID="f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" exitCode=0 Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831252 4796 generic.go:334] "Generic (PLEG): container finished" podID="a0383092-f637-4b82-a767-bf53ea59209e" containerID="371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" exitCode=0 Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831262 4796 generic.go:334] "Generic (PLEG): container finished" podID="a0383092-f637-4b82-a767-bf53ea59209e" containerID="4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" exitCode=0 Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831308 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerDied","Data":"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1"} Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831340 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerDied","Data":"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404"} Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831354 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerDied","Data":"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba"} Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831369 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a0383092-f637-4b82-a767-bf53ea59209e","Type":"ContainerDied","Data":"a5c352651f14edbf93ea78ea2cddc21cd33338540831229f48bc6a246e146772"} Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831388 4796 scope.go:117] "RemoveContainer" containerID="f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.831558 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.866167 4796 scope.go:117] "RemoveContainer" containerID="371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.899164 4796 scope.go:117] "RemoveContainer" containerID="4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.923576 4796 scope.go:117] "RemoveContainer" containerID="86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.946082 4796 scope.go:117] "RemoveContainer" containerID="f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" Nov 27 12:34:00 crc kubenswrapper[4796]: E1127 12:34:00.946647 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": container with ID starting with f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1 not found: ID does not exist" containerID="f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.946689 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1"} err="failed to get container status \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": rpc error: code = NotFound desc = could not find container \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": container with ID starting with f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.946714 4796 scope.go:117] "RemoveContainer" containerID="371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" Nov 27 12:34:00 crc kubenswrapper[4796]: E1127 12:34:00.947135 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": container with ID starting with 371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404 not found: ID does not exist" containerID="371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.947159 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404"} err="failed to get container status \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": rpc error: code = NotFound desc = could not find container \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": container with ID starting with 371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.947174 4796 scope.go:117] "RemoveContainer" containerID="4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" Nov 27 12:34:00 crc kubenswrapper[4796]: E1127 12:34:00.947956 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": container with ID starting with 4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba not found: ID does not exist" containerID="4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.947993 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba"} err="failed to get container status \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": rpc error: code = NotFound desc = could not find container \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": container with ID starting with 4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948022 4796 scope.go:117] "RemoveContainer" containerID="86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3" Nov 27 12:34:00 crc kubenswrapper[4796]: E1127 12:34:00.948296 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": container with ID starting with 86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3 not found: ID does not exist" containerID="86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948329 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3"} err="failed to get container status \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": rpc error: code = NotFound desc = could not find container \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": container with ID starting with 86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948347 4796 scope.go:117] "RemoveContainer" containerID="f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948564 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1"} err="failed to get container status \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": rpc error: code = NotFound desc = could not find container \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": container with ID starting with f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948589 4796 scope.go:117] "RemoveContainer" containerID="371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948806 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404"} err="failed to get container status \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": rpc error: code = NotFound desc = could not find container \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": container with ID starting with 371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.948829 4796 scope.go:117] "RemoveContainer" containerID="4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.949016 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba"} err="failed to get container status \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": rpc error: code = NotFound desc = could not find container \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": container with ID starting with 4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.949047 4796 scope.go:117] "RemoveContainer" containerID="86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.949376 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3"} err="failed to get container status \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": rpc error: code = NotFound desc = could not find container \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": container with ID starting with 86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.949401 4796 scope.go:117] "RemoveContainer" containerID="f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950081 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1"} err="failed to get container status \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": rpc error: code = NotFound desc = could not find container \"f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1\": container with ID starting with f60b83b9fd8ffdee7fbdbd8affd53c3b808d278708dc95109416d59308c8b3e1 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950114 4796 scope.go:117] "RemoveContainer" containerID="371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950432 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404"} err="failed to get container status \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": rpc error: code = NotFound desc = could not find container \"371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404\": container with ID starting with 371b44c4c15bf0175e02a524828bea4cde2b86124e1b72a1e398d32bc8f9b404 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950463 4796 scope.go:117] "RemoveContainer" containerID="4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950689 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba"} err="failed to get container status \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": rpc error: code = NotFound desc = could not find container \"4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba\": container with ID starting with 4a644a0b6b71a99e03798d00ed0bba2ab0466de282818233f30e632f2d6316ba not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950717 4796 scope.go:117] "RemoveContainer" containerID="86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.950984 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3"} err="failed to get container status \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": rpc error: code = NotFound desc = could not find container \"86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3\": container with ID starting with 86e45959bd2d6d853291819fcdc4f49af162abf7d509968108471a266fb0a7d3 not found: ID does not exist" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.967803 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vchl\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-kube-api-access-9vchl\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.967921 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-thanos-prometheus-http-client-file\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.967952 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0383092-f637-4b82-a767-bf53ea59209e-config-out\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.967974 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a0383092-f637-4b82-a767-bf53ea59209e-prometheus-metric-storage-rulefiles-0\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968179 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968285 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-config\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968347 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968376 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-secret-combined-ca-bundle\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968426 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968468 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.968503 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-tls-assets\") pod \"a0383092-f637-4b82-a767-bf53ea59209e\" (UID: \"a0383092-f637-4b82-a767-bf53ea59209e\") " Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.969235 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0383092-f637-4b82-a767-bf53ea59209e-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.975051 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0383092-f637-4b82-a767-bf53ea59209e-config-out" (OuterVolumeSpecName: "config-out") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.975083 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-kube-api-access-9vchl" (OuterVolumeSpecName: "kube-api-access-9vchl") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "kube-api-access-9vchl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.975091 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.977337 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.977417 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.977432 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.977597 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.977920 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-config" (OuterVolumeSpecName: "config") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:34:00 crc kubenswrapper[4796]: I1127 12:34:00.989793 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.058484 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config" (OuterVolumeSpecName: "web-config") pod "a0383092-f637-4b82-a767-bf53ea59209e" (UID: "a0383092-f637-4b82-a767-bf53ea59209e"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071258 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071302 4796 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071312 4796 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071323 4796 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071334 4796 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071345 4796 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071354 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vchl\" (UniqueName: \"kubernetes.io/projected/a0383092-f637-4b82-a767-bf53ea59209e-kube-api-access-9vchl\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071362 4796 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a0383092-f637-4b82-a767-bf53ea59209e-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071370 4796 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0383092-f637-4b82-a767-bf53ea59209e-config-out\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071379 4796 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a0383092-f637-4b82-a767-bf53ea59209e-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.071410 4796 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") on node \"crc\" " Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.092898 4796 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.093030 4796 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9") on node "crc" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.174312 4796 reconciler_common.go:293] "Volume detached for volume \"pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19e4c842-e39f-463a-8e01-73b3ecec16f9\") on node \"crc\" DevicePath \"\"" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.178823 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.195039 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.568755 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:34:01 crc kubenswrapper[4796]: E1127 12:34:01.569168 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:34:01 crc kubenswrapper[4796]: I1127 12:34:01.580849 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0383092-f637-4b82-a767-bf53ea59209e" path="/var/lib/kubelet/pods/a0383092-f637-4b82-a767-bf53ea59209e/volumes" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.147160 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.147934 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="config-reloader" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.147952 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="config-reloader" Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.147966 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="init-config-reloader" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.147974 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="init-config-reloader" Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.147993 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="registry-server" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148001 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="registry-server" Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.148019 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="prometheus" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148027 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="prometheus" Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.148047 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="extract-content" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148055 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="extract-content" Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.148075 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="extract-utilities" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148082 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="extract-utilities" Nov 27 12:34:02 crc kubenswrapper[4796]: E1127 12:34:02.148094 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="thanos-sidecar" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148101 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="thanos-sidecar" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148404 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ebbefc-c2f7-4c6d-9442-958d387621f0" containerName="registry-server" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148426 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="prometheus" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148443 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="thanos-sidecar" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.148460 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0383092-f637-4b82-a767-bf53ea59209e" containerName="config-reloader" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.150400 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.152437 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.153327 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.153650 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.153783 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.157237 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.159105 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6tv95" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.179695 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.191191 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210792 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210840 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210864 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210888 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210915 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210933 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.210962 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hns84\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-kube-api-access-hns84\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.211001 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.211050 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-config\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.211077 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312454 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312528 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hns84\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-kube-api-access-hns84\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312601 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312643 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-config\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312673 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312719 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312742 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312781 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312805 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312829 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.312858 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.313466 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.313747 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.317159 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.317508 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.318562 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.319299 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.319341 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.328778 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-config\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.329718 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.330198 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.334380 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hns84\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-kube-api-access-hns84\") pod \"prometheus-metric-storage-0\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.495869 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:02 crc kubenswrapper[4796]: I1127 12:34:02.956794 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:34:03 crc kubenswrapper[4796]: I1127 12:34:03.861448 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerStarted","Data":"0c3deee79ecc3b9d762100787aa6417dea8a803529a9f989c4a31453f3d724bc"} Nov 27 12:34:06 crc kubenswrapper[4796]: I1127 12:34:06.893507 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerStarted","Data":"1cc91c5a774d50a45d193b6b1c20ed2ae2817261a79a0cc2b5e4bc769707a946"} Nov 27 12:34:13 crc kubenswrapper[4796]: I1127 12:34:13.957995 4796 generic.go:334] "Generic (PLEG): container finished" podID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerID="1cc91c5a774d50a45d193b6b1c20ed2ae2817261a79a0cc2b5e4bc769707a946" exitCode=0 Nov 27 12:34:13 crc kubenswrapper[4796]: I1127 12:34:13.958089 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerDied","Data":"1cc91c5a774d50a45d193b6b1c20ed2ae2817261a79a0cc2b5e4bc769707a946"} Nov 27 12:34:14 crc kubenswrapper[4796]: I1127 12:34:14.569442 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:34:14 crc kubenswrapper[4796]: E1127 12:34:14.570209 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:34:14 crc kubenswrapper[4796]: I1127 12:34:14.972310 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerStarted","Data":"33e1fe9f5eb69f50390441e19c57b6309eff3a914eb21df917e7e089194f53f5"} Nov 27 12:34:19 crc kubenswrapper[4796]: I1127 12:34:19.036996 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerStarted","Data":"0741db1ad66b0c7658c7ca79059c06368d5d9f5f954487584311ad9229bc3896"} Nov 27 12:34:19 crc kubenswrapper[4796]: I1127 12:34:19.037454 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerStarted","Data":"4457425367e16c0e2cab63db27dd778f76f8da200ce0661aba64c67a2679de90"} Nov 27 12:34:19 crc kubenswrapper[4796]: I1127 12:34:19.063636 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.063618197 podStartE2EDuration="17.063618197s" podCreationTimestamp="2025-11-27 12:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:34:19.063057972 +0000 UTC m=+4176.581376890" watchObservedRunningTime="2025-11-27 12:34:19.063618197 +0000 UTC m=+4176.581937115" Nov 27 12:34:22 crc kubenswrapper[4796]: I1127 12:34:22.496937 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:29 crc kubenswrapper[4796]: I1127 12:34:29.569684 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:34:29 crc kubenswrapper[4796]: E1127 12:34:29.570479 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:34:32 crc kubenswrapper[4796]: I1127 12:34:32.496078 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:32 crc kubenswrapper[4796]: I1127 12:34:32.503642 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:33 crc kubenswrapper[4796]: I1127 12:34:33.181741 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 27 12:34:41 crc kubenswrapper[4796]: I1127 12:34:41.569098 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:34:42 crc kubenswrapper[4796]: I1127 12:34:42.284623 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"2c9f964331a818e21cdd07be85bc6c8a01ad0977a6954ef3cab78c4971adbb4e"} Nov 27 12:35:43 crc kubenswrapper[4796]: I1127 12:35:43.051163 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-nbbj5"] Nov 27 12:35:43 crc kubenswrapper[4796]: I1127 12:35:43.060264 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-cbbf-account-create-update-bwk2h"] Nov 27 12:35:43 crc kubenswrapper[4796]: I1127 12:35:43.071632 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-cbbf-account-create-update-bwk2h"] Nov 27 12:35:43 crc kubenswrapper[4796]: I1127 12:35:43.079566 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-nbbj5"] Nov 27 12:35:43 crc kubenswrapper[4796]: I1127 12:35:43.585386 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab498b8b-bcf0-4e08-82a9-ec116273269c" path="/var/lib/kubelet/pods/ab498b8b-bcf0-4e08-82a9-ec116273269c/volumes" Nov 27 12:35:43 crc kubenswrapper[4796]: I1127 12:35:43.587596 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afe8e4fe-1791-461c-a9bc-bb9bd352d7d9" path="/var/lib/kubelet/pods/afe8e4fe-1791-461c-a9bc-bb9bd352d7d9/volumes" Nov 27 12:35:54 crc kubenswrapper[4796]: I1127 12:35:54.216248 4796 scope.go:117] "RemoveContainer" containerID="263ba999cfc7e9c5722643ea26eeef99004d6d67e580a46ddcab4bc20c2b482d" Nov 27 12:35:54 crc kubenswrapper[4796]: I1127 12:35:54.252123 4796 scope.go:117] "RemoveContainer" containerID="59019cb781f6b79431e6ad89679a8a559b4d315656201884fda1e550b20a5d14" Nov 27 12:35:55 crc kubenswrapper[4796]: I1127 12:35:55.044974 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-rhjzb"] Nov 27 12:35:55 crc kubenswrapper[4796]: I1127 12:35:55.052684 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-rhjzb"] Nov 27 12:35:55 crc kubenswrapper[4796]: I1127 12:35:55.591135 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7aba55-765f-4264-b1c7-803655588e9a" path="/var/lib/kubelet/pods/bb7aba55-765f-4264-b1c7-803655588e9a/volumes" Nov 27 12:35:59 crc kubenswrapper[4796]: I1127 12:35:59.997765 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:36:01 crc kubenswrapper[4796]: I1127 12:36:01.421003 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:36:01 crc kubenswrapper[4796]: I1127 12:36:01.422250 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-api" containerID="cri-o://39aaaf3c94179a85c88b3e5bba465336efc123996659a405bfaae7b9f2278475" gracePeriod=30 Nov 27 12:36:01 crc kubenswrapper[4796]: I1127 12:36:01.422459 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-evaluator" containerID="cri-o://5d576b837960794610df67aa6caa98c973f88e660b4def1ab1732453d33a5b2f" gracePeriod=30 Nov 27 12:36:01 crc kubenswrapper[4796]: I1127 12:36:01.422303 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-notifier" containerID="cri-o://a4e657b598b67d12de4572a98206413e70744b00b8583ed9b8dab3a970415cfa" gracePeriod=30 Nov 27 12:36:01 crc kubenswrapper[4796]: I1127 12:36:01.422294 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-listener" containerID="cri-o://b9fe67225938676744c90f71d643031ffec9ceafd68ca980dff4391ec99f9a76" gracePeriod=30 Nov 27 12:36:02 crc kubenswrapper[4796]: I1127 12:36:02.074230 4796 generic.go:334] "Generic (PLEG): container finished" podID="220303ef-0c3c-4c74-959a-129565095e51" containerID="5d576b837960794610df67aa6caa98c973f88e660b4def1ab1732453d33a5b2f" exitCode=0 Nov 27 12:36:02 crc kubenswrapper[4796]: I1127 12:36:02.075326 4796 generic.go:334] "Generic (PLEG): container finished" podID="220303ef-0c3c-4c74-959a-129565095e51" containerID="39aaaf3c94179a85c88b3e5bba465336efc123996659a405bfaae7b9f2278475" exitCode=0 Nov 27 12:36:02 crc kubenswrapper[4796]: I1127 12:36:02.075039 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerDied","Data":"5d576b837960794610df67aa6caa98c973f88e660b4def1ab1732453d33a5b2f"} Nov 27 12:36:02 crc kubenswrapper[4796]: I1127 12:36:02.076092 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerDied","Data":"39aaaf3c94179a85c88b3e5bba465336efc123996659a405bfaae7b9f2278475"} Nov 27 12:36:09 crc kubenswrapper[4796]: I1127 12:36:09.143295 4796 generic.go:334] "Generic (PLEG): container finished" podID="220303ef-0c3c-4c74-959a-129565095e51" containerID="a4e657b598b67d12de4572a98206413e70744b00b8583ed9b8dab3a970415cfa" exitCode=0 Nov 27 12:36:09 crc kubenswrapper[4796]: I1127 12:36:09.143372 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerDied","Data":"a4e657b598b67d12de4572a98206413e70744b00b8583ed9b8dab3a970415cfa"} Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.158622 4796 generic.go:334] "Generic (PLEG): container finished" podID="220303ef-0c3c-4c74-959a-129565095e51" containerID="b9fe67225938676744c90f71d643031ffec9ceafd68ca980dff4391ec99f9a76" exitCode=0 Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.158677 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerDied","Data":"b9fe67225938676744c90f71d643031ffec9ceafd68ca980dff4391ec99f9a76"} Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.664598 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.678109 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-scripts\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.678248 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-internal-tls-certs\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.678364 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-config-data\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.678401 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.678495 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-combined-ca-bundle\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.678621 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/220303ef-0c3c-4c74-959a-129565095e51-kube-api-access-vchhk\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.684079 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-scripts" (OuterVolumeSpecName: "scripts") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.699514 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/220303ef-0c3c-4c74-959a-129565095e51-kube-api-access-vchhk" (OuterVolumeSpecName: "kube-api-access-vchhk") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "kube-api-access-vchhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.757447 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.781821 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.782315 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs\") pod \"220303ef-0c3c-4c74-959a-129565095e51\" (UID: \"220303ef-0c3c-4c74-959a-129565095e51\") " Nov 27 12:36:10 crc kubenswrapper[4796]: W1127 12:36:10.782463 4796 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/220303ef-0c3c-4c74-959a-129565095e51/volumes/kubernetes.io~secret/public-tls-certs Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.782476 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.782905 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/220303ef-0c3c-4c74-959a-129565095e51-kube-api-access-vchhk\") on node \"crc\" DevicePath \"\"" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.782918 4796 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.782927 4796 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.782937 4796 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.825476 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-config-data" (OuterVolumeSpecName: "config-data") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.849468 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "220303ef-0c3c-4c74-959a-129565095e51" (UID: "220303ef-0c3c-4c74-959a-129565095e51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.885421 4796 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 12:36:10 crc kubenswrapper[4796]: I1127 12:36:10.885460 4796 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220303ef-0c3c-4c74-959a-129565095e51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.172099 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"220303ef-0c3c-4c74-959a-129565095e51","Type":"ContainerDied","Data":"ec75cc995c6c01f9624ea93cd210295df643cd96af2d8d02e8a675b67d4002d3"} Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.172183 4796 scope.go:117] "RemoveContainer" containerID="b9fe67225938676744c90f71d643031ffec9ceafd68ca980dff4391ec99f9a76" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.172221 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.226545 4796 scope.go:117] "RemoveContainer" containerID="a4e657b598b67d12de4572a98206413e70744b00b8583ed9b8dab3a970415cfa" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.245652 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.261588 4796 scope.go:117] "RemoveContainer" containerID="5d576b837960794610df67aa6caa98c973f88e660b4def1ab1732453d33a5b2f" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.262186 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.270629 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 27 12:36:11 crc kubenswrapper[4796]: E1127 12:36:11.271015 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-evaluator" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271031 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-evaluator" Nov 27 12:36:11 crc kubenswrapper[4796]: E1127 12:36:11.271046 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-notifier" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271053 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-notifier" Nov 27 12:36:11 crc kubenswrapper[4796]: E1127 12:36:11.271066 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-api" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271072 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-api" Nov 27 12:36:11 crc kubenswrapper[4796]: E1127 12:36:11.271082 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-listener" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271087 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-listener" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271255 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-notifier" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271284 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-evaluator" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271298 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-api" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.271312 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="220303ef-0c3c-4c74-959a-129565095e51" containerName="aodh-listener" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.273211 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.275938 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.276254 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vs2ld" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.276404 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.276657 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.276834 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.286949 4796 scope.go:117] "RemoveContainer" containerID="39aaaf3c94179a85c88b3e5bba465336efc123996659a405bfaae7b9f2278475" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.293963 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.294018 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-internal-tls-certs\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.294065 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-public-tls-certs\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.294081 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp5fz\" (UniqueName: \"kubernetes.io/projected/85e2137f-a13a-480b-9d77-e2bafab58d5d-kube-api-access-bp5fz\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.294104 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-config-data\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.294139 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-scripts\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.302369 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.395705 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.395792 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-internal-tls-certs\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.395850 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-public-tls-certs\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.395876 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp5fz\" (UniqueName: \"kubernetes.io/projected/85e2137f-a13a-480b-9d77-e2bafab58d5d-kube-api-access-bp5fz\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.395904 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-config-data\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.395967 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-scripts\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.401218 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.401249 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-internal-tls-certs\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.401219 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-scripts\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.401951 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-public-tls-certs\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.408042 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e2137f-a13a-480b-9d77-e2bafab58d5d-config-data\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.421530 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp5fz\" (UniqueName: \"kubernetes.io/projected/85e2137f-a13a-480b-9d77-e2bafab58d5d-kube-api-access-bp5fz\") pod \"aodh-0\" (UID: \"85e2137f-a13a-480b-9d77-e2bafab58d5d\") " pod="openstack/aodh-0" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.580560 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="220303ef-0c3c-4c74-959a-129565095e51" path="/var/lib/kubelet/pods/220303ef-0c3c-4c74-959a-129565095e51/volumes" Nov 27 12:36:11 crc kubenswrapper[4796]: I1127 12:36:11.592086 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 27 12:36:12 crc kubenswrapper[4796]: I1127 12:36:12.057625 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 27 12:36:12 crc kubenswrapper[4796]: I1127 12:36:12.067521 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:36:12 crc kubenswrapper[4796]: I1127 12:36:12.183452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85e2137f-a13a-480b-9d77-e2bafab58d5d","Type":"ContainerStarted","Data":"bd9bc6844b01b2bd6b1d3ccaa2af3f82562e5253f938dfaaf4012008a293fe43"} Nov 27 12:36:13 crc kubenswrapper[4796]: I1127 12:36:13.194452 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85e2137f-a13a-480b-9d77-e2bafab58d5d","Type":"ContainerStarted","Data":"ad9be5b76d0a11b373022e4e635af4a7c9441c706d60262961dfb16fad7c5fbc"} Nov 27 12:36:14 crc kubenswrapper[4796]: I1127 12:36:14.207255 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85e2137f-a13a-480b-9d77-e2bafab58d5d","Type":"ContainerStarted","Data":"ed8421572141404154a077dd6b06f4cc2e1f1f286cfe211491c401dfb288d6d3"} Nov 27 12:36:15 crc kubenswrapper[4796]: I1127 12:36:15.217790 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85e2137f-a13a-480b-9d77-e2bafab58d5d","Type":"ContainerStarted","Data":"68a3fc8294fd7d3d4e96f4a4b7dcd87317b6ff256e0a177e55424c2fc185e68e"} Nov 27 12:36:16 crc kubenswrapper[4796]: I1127 12:36:16.228500 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85e2137f-a13a-480b-9d77-e2bafab58d5d","Type":"ContainerStarted","Data":"85fe3b98348ee3c85c25fa424bbcc57cee2b25f87e02d6cf14f748eab6eb31ac"} Nov 27 12:36:16 crc kubenswrapper[4796]: I1127 12:36:16.263711 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.929151821 podStartE2EDuration="5.263689548s" podCreationTimestamp="2025-11-27 12:36:11 +0000 UTC" firstStartedPulling="2025-11-27 12:36:12.067208072 +0000 UTC m=+4289.585526990" lastFinishedPulling="2025-11-27 12:36:15.401745799 +0000 UTC m=+4292.920064717" observedRunningTime="2025-11-27 12:36:16.24538844 +0000 UTC m=+4293.763707388" watchObservedRunningTime="2025-11-27 12:36:16.263689548 +0000 UTC m=+4293.782008466" Nov 27 12:36:54 crc kubenswrapper[4796]: I1127 12:36:54.398029 4796 scope.go:117] "RemoveContainer" containerID="9ed5b99cec6c084e247a7bba7eccfd000c31e8c094e4b7282a667c934712bce2" Nov 27 12:37:01 crc kubenswrapper[4796]: I1127 12:37:01.879220 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:37:01 crc kubenswrapper[4796]: I1127 12:37:01.879874 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.461305 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mrc2n"] Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.465669 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.486493 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mrc2n"] Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.561905 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-utilities\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.561971 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-catalog-content\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.562038 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdvm\" (UniqueName: \"kubernetes.io/projected/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-kube-api-access-xrdvm\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.663408 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdvm\" (UniqueName: \"kubernetes.io/projected/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-kube-api-access-xrdvm\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.663628 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-utilities\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.663673 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-catalog-content\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.664116 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-utilities\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.664165 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-catalog-content\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.685128 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdvm\" (UniqueName: \"kubernetes.io/projected/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-kube-api-access-xrdvm\") pod \"redhat-operators-mrc2n\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:05 crc kubenswrapper[4796]: I1127 12:37:05.799962 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:06 crc kubenswrapper[4796]: I1127 12:37:06.287679 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mrc2n"] Nov 27 12:37:06 crc kubenswrapper[4796]: I1127 12:37:06.720300 4796 generic.go:334] "Generic (PLEG): container finished" podID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerID="4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4" exitCode=0 Nov 27 12:37:06 crc kubenswrapper[4796]: I1127 12:37:06.720391 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerDied","Data":"4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4"} Nov 27 12:37:06 crc kubenswrapper[4796]: I1127 12:37:06.720537 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerStarted","Data":"b8944805bc4f8e25d09aa9095789e92d3ec7426ccb2c15b01e850734fcf7246a"} Nov 27 12:37:07 crc kubenswrapper[4796]: I1127 12:37:07.733729 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerStarted","Data":"cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b"} Nov 27 12:37:12 crc kubenswrapper[4796]: I1127 12:37:12.785369 4796 generic.go:334] "Generic (PLEG): container finished" podID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerID="cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b" exitCode=0 Nov 27 12:37:12 crc kubenswrapper[4796]: I1127 12:37:12.785427 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerDied","Data":"cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b"} Nov 27 12:37:13 crc kubenswrapper[4796]: I1127 12:37:13.800549 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerStarted","Data":"c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d"} Nov 27 12:37:13 crc kubenswrapper[4796]: I1127 12:37:13.829620 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mrc2n" podStartSLOduration=2.033929317 podStartE2EDuration="8.829594835s" podCreationTimestamp="2025-11-27 12:37:05 +0000 UTC" firstStartedPulling="2025-11-27 12:37:06.722125305 +0000 UTC m=+4344.240444213" lastFinishedPulling="2025-11-27 12:37:13.517790813 +0000 UTC m=+4351.036109731" observedRunningTime="2025-11-27 12:37:13.81837282 +0000 UTC m=+4351.336691748" watchObservedRunningTime="2025-11-27 12:37:13.829594835 +0000 UTC m=+4351.347913773" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.447764 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t22sd"] Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.450217 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.468548 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t22sd"] Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.635462 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-utilities\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.635609 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xkdg\" (UniqueName: \"kubernetes.io/projected/78480faa-f85a-4a68-a288-5a208041dd98-kube-api-access-9xkdg\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.635905 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-catalog-content\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.738641 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xkdg\" (UniqueName: \"kubernetes.io/projected/78480faa-f85a-4a68-a288-5a208041dd98-kube-api-access-9xkdg\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.738709 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-utilities\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.738958 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-catalog-content\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.739617 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-utilities\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.739751 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-catalog-content\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.761883 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xkdg\" (UniqueName: \"kubernetes.io/projected/78480faa-f85a-4a68-a288-5a208041dd98-kube-api-access-9xkdg\") pod \"redhat-marketplace-t22sd\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:14 crc kubenswrapper[4796]: I1127 12:37:14.782506 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:15 crc kubenswrapper[4796]: I1127 12:37:15.223552 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t22sd"] Nov 27 12:37:15 crc kubenswrapper[4796]: I1127 12:37:15.800225 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:15 crc kubenswrapper[4796]: I1127 12:37:15.800283 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:15 crc kubenswrapper[4796]: I1127 12:37:15.820435 4796 generic.go:334] "Generic (PLEG): container finished" podID="78480faa-f85a-4a68-a288-5a208041dd98" containerID="b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc" exitCode=0 Nov 27 12:37:15 crc kubenswrapper[4796]: I1127 12:37:15.821033 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerDied","Data":"b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc"} Nov 27 12:37:15 crc kubenswrapper[4796]: I1127 12:37:15.821219 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerStarted","Data":"91307582aef5a99a31a447aa2ece8cd75c6ff20a463cfcf19cefb181f11b218a"} Nov 27 12:37:16 crc kubenswrapper[4796]: I1127 12:37:16.831648 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerStarted","Data":"86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2"} Nov 27 12:37:16 crc kubenswrapper[4796]: I1127 12:37:16.852100 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mrc2n" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="registry-server" probeResult="failure" output=< Nov 27 12:37:16 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:37:16 crc kubenswrapper[4796]: > Nov 27 12:37:17 crc kubenswrapper[4796]: I1127 12:37:17.842714 4796 generic.go:334] "Generic (PLEG): container finished" podID="78480faa-f85a-4a68-a288-5a208041dd98" containerID="86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2" exitCode=0 Nov 27 12:37:17 crc kubenswrapper[4796]: I1127 12:37:17.842782 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerDied","Data":"86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2"} Nov 27 12:37:18 crc kubenswrapper[4796]: I1127 12:37:18.865017 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerStarted","Data":"9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8"} Nov 27 12:37:18 crc kubenswrapper[4796]: I1127 12:37:18.890430 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t22sd" podStartSLOduration=2.381304215 podStartE2EDuration="4.890411925s" podCreationTimestamp="2025-11-27 12:37:14 +0000 UTC" firstStartedPulling="2025-11-27 12:37:15.822303867 +0000 UTC m=+4353.340622785" lastFinishedPulling="2025-11-27 12:37:18.331411577 +0000 UTC m=+4355.849730495" observedRunningTime="2025-11-27 12:37:18.882565111 +0000 UTC m=+4356.400884029" watchObservedRunningTime="2025-11-27 12:37:18.890411925 +0000 UTC m=+4356.408730843" Nov 27 12:37:25 crc kubenswrapper[4796]: I1127 12:37:25.005755 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:25 crc kubenswrapper[4796]: I1127 12:37:25.007107 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:25 crc kubenswrapper[4796]: I1127 12:37:25.066453 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:26 crc kubenswrapper[4796]: I1127 12:37:26.078557 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:26 crc kubenswrapper[4796]: I1127 12:37:26.127411 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t22sd"] Nov 27 12:37:26 crc kubenswrapper[4796]: I1127 12:37:26.852412 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mrc2n" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="registry-server" probeResult="failure" output=< Nov 27 12:37:26 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:37:26 crc kubenswrapper[4796]: > Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.041181 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t22sd" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="registry-server" containerID="cri-o://9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8" gracePeriod=2 Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.508189 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.638610 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-catalog-content\") pod \"78480faa-f85a-4a68-a288-5a208041dd98\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.638973 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xkdg\" (UniqueName: \"kubernetes.io/projected/78480faa-f85a-4a68-a288-5a208041dd98-kube-api-access-9xkdg\") pod \"78480faa-f85a-4a68-a288-5a208041dd98\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.639161 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-utilities\") pod \"78480faa-f85a-4a68-a288-5a208041dd98\" (UID: \"78480faa-f85a-4a68-a288-5a208041dd98\") " Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.639829 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-utilities" (OuterVolumeSpecName: "utilities") pod "78480faa-f85a-4a68-a288-5a208041dd98" (UID: "78480faa-f85a-4a68-a288-5a208041dd98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.645099 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78480faa-f85a-4a68-a288-5a208041dd98-kube-api-access-9xkdg" (OuterVolumeSpecName: "kube-api-access-9xkdg") pod "78480faa-f85a-4a68-a288-5a208041dd98" (UID: "78480faa-f85a-4a68-a288-5a208041dd98"). InnerVolumeSpecName "kube-api-access-9xkdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.657309 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78480faa-f85a-4a68-a288-5a208041dd98" (UID: "78480faa-f85a-4a68-a288-5a208041dd98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.741783 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.741825 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78480faa-f85a-4a68-a288-5a208041dd98-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:37:28 crc kubenswrapper[4796]: I1127 12:37:28.741842 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xkdg\" (UniqueName: \"kubernetes.io/projected/78480faa-f85a-4a68-a288-5a208041dd98-kube-api-access-9xkdg\") on node \"crc\" DevicePath \"\"" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.061283 4796 generic.go:334] "Generic (PLEG): container finished" podID="78480faa-f85a-4a68-a288-5a208041dd98" containerID="9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8" exitCode=0 Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.061337 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerDied","Data":"9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8"} Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.061371 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t22sd" event={"ID":"78480faa-f85a-4a68-a288-5a208041dd98","Type":"ContainerDied","Data":"91307582aef5a99a31a447aa2ece8cd75c6ff20a463cfcf19cefb181f11b218a"} Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.061392 4796 scope.go:117] "RemoveContainer" containerID="9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.061903 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t22sd" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.099394 4796 scope.go:117] "RemoveContainer" containerID="86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.123428 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t22sd"] Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.138210 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t22sd"] Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.140571 4796 scope.go:117] "RemoveContainer" containerID="b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.190529 4796 scope.go:117] "RemoveContainer" containerID="9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8" Nov 27 12:37:29 crc kubenswrapper[4796]: E1127 12:37:29.190938 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8\": container with ID starting with 9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8 not found: ID does not exist" containerID="9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.190979 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8"} err="failed to get container status \"9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8\": rpc error: code = NotFound desc = could not find container \"9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8\": container with ID starting with 9b77b65d901b80f2d2a281897cc56b87f53198383037feaeefe20d4bcfc115f8 not found: ID does not exist" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.191005 4796 scope.go:117] "RemoveContainer" containerID="86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2" Nov 27 12:37:29 crc kubenswrapper[4796]: E1127 12:37:29.191720 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2\": container with ID starting with 86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2 not found: ID does not exist" containerID="86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.191797 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2"} err="failed to get container status \"86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2\": rpc error: code = NotFound desc = could not find container \"86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2\": container with ID starting with 86334d0e41fc7dfe389bf4eb636e773226e7d9d2b1feb992fce789a9c9f8eec2 not found: ID does not exist" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.191859 4796 scope.go:117] "RemoveContainer" containerID="b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc" Nov 27 12:37:29 crc kubenswrapper[4796]: E1127 12:37:29.192247 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc\": container with ID starting with b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc not found: ID does not exist" containerID="b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.192360 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc"} err="failed to get container status \"b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc\": rpc error: code = NotFound desc = could not find container \"b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc\": container with ID starting with b8fa9063f5f9ba3ee55219ffdfaa7719c6ba8a9e0e1409b4b873b93db48997fc not found: ID does not exist" Nov 27 12:37:29 crc kubenswrapper[4796]: I1127 12:37:29.585431 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78480faa-f85a-4a68-a288-5a208041dd98" path="/var/lib/kubelet/pods/78480faa-f85a-4a68-a288-5a208041dd98/volumes" Nov 27 12:37:31 crc kubenswrapper[4796]: I1127 12:37:31.879511 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:37:31 crc kubenswrapper[4796]: I1127 12:37:31.880083 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:37:35 crc kubenswrapper[4796]: I1127 12:37:35.858755 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:35 crc kubenswrapper[4796]: I1127 12:37:35.925163 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:36 crc kubenswrapper[4796]: I1127 12:37:36.660705 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mrc2n"] Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.136779 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mrc2n" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="registry-server" containerID="cri-o://c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d" gracePeriod=2 Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.610592 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.715142 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-utilities\") pod \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.715236 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-catalog-content\") pod \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.715359 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrdvm\" (UniqueName: \"kubernetes.io/projected/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-kube-api-access-xrdvm\") pod \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\" (UID: \"42f2d64b-e1f9-4702-b467-0ace2ab3fe36\") " Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.716728 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-utilities" (OuterVolumeSpecName: "utilities") pod "42f2d64b-e1f9-4702-b467-0ace2ab3fe36" (UID: "42f2d64b-e1f9-4702-b467-0ace2ab3fe36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.722045 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-kube-api-access-xrdvm" (OuterVolumeSpecName: "kube-api-access-xrdvm") pod "42f2d64b-e1f9-4702-b467-0ace2ab3fe36" (UID: "42f2d64b-e1f9-4702-b467-0ace2ab3fe36"). InnerVolumeSpecName "kube-api-access-xrdvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.808884 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42f2d64b-e1f9-4702-b467-0ace2ab3fe36" (UID: "42f2d64b-e1f9-4702-b467-0ace2ab3fe36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.818158 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.818195 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:37:37 crc kubenswrapper[4796]: I1127 12:37:37.818209 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrdvm\" (UniqueName: \"kubernetes.io/projected/42f2d64b-e1f9-4702-b467-0ace2ab3fe36-kube-api-access-xrdvm\") on node \"crc\" DevicePath \"\"" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.149689 4796 generic.go:334] "Generic (PLEG): container finished" podID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerID="c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d" exitCode=0 Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.149751 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerDied","Data":"c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d"} Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.149786 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mrc2n" event={"ID":"42f2d64b-e1f9-4702-b467-0ace2ab3fe36","Type":"ContainerDied","Data":"b8944805bc4f8e25d09aa9095789e92d3ec7426ccb2c15b01e850734fcf7246a"} Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.149808 4796 scope.go:117] "RemoveContainer" containerID="c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.149864 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mrc2n" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.187719 4796 scope.go:117] "RemoveContainer" containerID="cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.205922 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mrc2n"] Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.214074 4796 scope.go:117] "RemoveContainer" containerID="4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.214697 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mrc2n"] Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.264821 4796 scope.go:117] "RemoveContainer" containerID="c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d" Nov 27 12:37:38 crc kubenswrapper[4796]: E1127 12:37:38.265866 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d\": container with ID starting with c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d not found: ID does not exist" containerID="c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.265924 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d"} err="failed to get container status \"c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d\": rpc error: code = NotFound desc = could not find container \"c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d\": container with ID starting with c6dbbd2e6b55100d65b53678a58f71cd9a8b623a06135a7c55647c853b2e0d0d not found: ID does not exist" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.265955 4796 scope.go:117] "RemoveContainer" containerID="cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b" Nov 27 12:37:38 crc kubenswrapper[4796]: E1127 12:37:38.266525 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b\": container with ID starting with cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b not found: ID does not exist" containerID="cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.266590 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b"} err="failed to get container status \"cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b\": rpc error: code = NotFound desc = could not find container \"cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b\": container with ID starting with cd4f291b8977a5bbe636b4f85311ce0ebc92ff64c49a30e1d06cf25bdc75756b not found: ID does not exist" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.266622 4796 scope.go:117] "RemoveContainer" containerID="4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4" Nov 27 12:37:38 crc kubenswrapper[4796]: E1127 12:37:38.266889 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4\": container with ID starting with 4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4 not found: ID does not exist" containerID="4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4" Nov 27 12:37:38 crc kubenswrapper[4796]: I1127 12:37:38.266940 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4"} err="failed to get container status \"4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4\": rpc error: code = NotFound desc = could not find container \"4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4\": container with ID starting with 4a2535b6bb142acc00941abccd42105b07bcc77a91dcd1022e7f18a39033fce4 not found: ID does not exist" Nov 27 12:37:39 crc kubenswrapper[4796]: I1127 12:37:39.579520 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" path="/var/lib/kubelet/pods/42f2d64b-e1f9-4702-b467-0ace2ab3fe36/volumes" Nov 27 12:38:01 crc kubenswrapper[4796]: I1127 12:38:01.879335 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:38:01 crc kubenswrapper[4796]: I1127 12:38:01.879992 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:38:01 crc kubenswrapper[4796]: I1127 12:38:01.880045 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:38:01 crc kubenswrapper[4796]: I1127 12:38:01.880953 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c9f964331a818e21cdd07be85bc6c8a01ad0977a6954ef3cab78c4971adbb4e"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:38:01 crc kubenswrapper[4796]: I1127 12:38:01.881036 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://2c9f964331a818e21cdd07be85bc6c8a01ad0977a6954ef3cab78c4971adbb4e" gracePeriod=600 Nov 27 12:38:01 crc kubenswrapper[4796]: I1127 12:38:01.900524 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:38:02 crc kubenswrapper[4796]: I1127 12:38:02.401064 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="2c9f964331a818e21cdd07be85bc6c8a01ad0977a6954ef3cab78c4971adbb4e" exitCode=0 Nov 27 12:38:02 crc kubenswrapper[4796]: I1127 12:38:02.401477 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"2c9f964331a818e21cdd07be85bc6c8a01ad0977a6954ef3cab78c4971adbb4e"} Nov 27 12:38:02 crc kubenswrapper[4796]: I1127 12:38:02.401824 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648"} Nov 27 12:38:02 crc kubenswrapper[4796]: I1127 12:38:02.401869 4796 scope.go:117] "RemoveContainer" containerID="3f663e831987891f7369068aa67f2ed933c268075098e5718a885e6330155a58" Nov 27 12:38:05 crc kubenswrapper[4796]: I1127 12:38:05.415966 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:38:05 crc kubenswrapper[4796]: I1127 12:38:05.416884 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="prometheus" containerID="cri-o://33e1fe9f5eb69f50390441e19c57b6309eff3a914eb21df917e7e089194f53f5" gracePeriod=600 Nov 27 12:38:05 crc kubenswrapper[4796]: I1127 12:38:05.417001 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="thanos-sidecar" containerID="cri-o://0741db1ad66b0c7658c7ca79059c06368d5d9f5f954487584311ad9229bc3896" gracePeriod=600 Nov 27 12:38:05 crc kubenswrapper[4796]: I1127 12:38:05.417014 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="config-reloader" containerID="cri-o://4457425367e16c0e2cab63db27dd778f76f8da200ce0661aba64c67a2679de90" gracePeriod=600 Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.442413 4796 generic.go:334] "Generic (PLEG): container finished" podID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerID="0741db1ad66b0c7658c7ca79059c06368d5d9f5f954487584311ad9229bc3896" exitCode=0 Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.442946 4796 generic.go:334] "Generic (PLEG): container finished" podID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerID="4457425367e16c0e2cab63db27dd778f76f8da200ce0661aba64c67a2679de90" exitCode=0 Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.442956 4796 generic.go:334] "Generic (PLEG): container finished" podID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerID="33e1fe9f5eb69f50390441e19c57b6309eff3a914eb21df917e7e089194f53f5" exitCode=0 Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.442490 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerDied","Data":"0741db1ad66b0c7658c7ca79059c06368d5d9f5f954487584311ad9229bc3896"} Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.442992 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerDied","Data":"4457425367e16c0e2cab63db27dd778f76f8da200ce0661aba64c67a2679de90"} Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.443006 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerDied","Data":"33e1fe9f5eb69f50390441e19c57b6309eff3a914eb21df917e7e089194f53f5"} Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.784837 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.892292 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-secret-combined-ca-bundle\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.892759 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-tls-assets\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.892842 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-config\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.892873 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-thanos-prometheus-http-client-file\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.892926 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.892974 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.893025 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-config-out\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.893106 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-db\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.893146 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.893200 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hns84\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-kube-api-access-hns84\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.893232 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-rulefiles-0\") pod \"d02b35fa-5c27-4902-99fe-5c7370e34690\" (UID: \"d02b35fa-5c27-4902-99fe-5c7370e34690\") " Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.894225 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.895650 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-db" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "prometheus-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.899457 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-config-out" (OuterVolumeSpecName: "config-out") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.899493 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.900095 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-config" (OuterVolumeSpecName: "config") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.900817 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.900821 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.901946 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.902161 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-kube-api-access-hns84" (OuterVolumeSpecName: "kube-api-access-hns84") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "kube-api-access-hns84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.902281 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.974874 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config" (OuterVolumeSpecName: "web-config") pod "d02b35fa-5c27-4902-99fe-5c7370e34690" (UID: "d02b35fa-5c27-4902-99fe-5c7370e34690"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.995179 4796 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.995220 4796 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.995235 4796 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.995247 4796 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.995259 4796 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-config-out\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.996447 4796 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-db\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.996744 4796 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-web-config\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.996777 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hns84\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-kube-api-access-hns84\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.996796 4796 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d02b35fa-5c27-4902-99fe-5c7370e34690-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.996810 4796 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d02b35fa-5c27-4902-99fe-5c7370e34690-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:06 crc kubenswrapper[4796]: I1127 12:38:06.996822 4796 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d02b35fa-5c27-4902-99fe-5c7370e34690-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.454579 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"d02b35fa-5c27-4902-99fe-5c7370e34690","Type":"ContainerDied","Data":"0c3deee79ecc3b9d762100787aa6417dea8a803529a9f989c4a31453f3d724bc"} Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.454628 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.454642 4796 scope.go:117] "RemoveContainer" containerID="0741db1ad66b0c7658c7ca79059c06368d5d9f5f954487584311ad9229bc3896" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.479180 4796 scope.go:117] "RemoveContainer" containerID="4457425367e16c0e2cab63db27dd778f76f8da200ce0661aba64c67a2679de90" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.491471 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.500116 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.509951 4796 scope.go:117] "RemoveContainer" containerID="33e1fe9f5eb69f50390441e19c57b6309eff3a914eb21df917e7e089194f53f5" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.512775 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.513234 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="prometheus" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.513329 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="prometheus" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.513402 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="extract-content" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.513462 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="extract-content" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.513526 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="extract-utilities" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.513575 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="extract-utilities" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.513642 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="init-config-reloader" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.513701 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="init-config-reloader" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.513763 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="thanos-sidecar" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.513820 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="thanos-sidecar" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.513887 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="registry-server" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.513942 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="registry-server" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.514002 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="extract-utilities" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514054 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="extract-utilities" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.514110 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="config-reloader" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514162 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="config-reloader" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.514215 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="extract-content" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514287 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="extract-content" Nov 27 12:38:07 crc kubenswrapper[4796]: E1127 12:38:07.514352 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="registry-server" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514401 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="registry-server" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514615 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="config-reloader" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514685 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="78480faa-f85a-4a68-a288-5a208041dd98" containerName="registry-server" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.514947 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="prometheus" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.515052 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f2d64b-e1f9-4702-b467-0ace2ab3fe36" containerName="registry-server" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.515352 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" containerName="thanos-sidecar" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.517707 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.520184 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.520389 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.521101 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.521171 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.521234 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.522527 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-6tv95" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.529088 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.534110 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.536448 4796 scope.go:117] "RemoveContainer" containerID="1cc91c5a774d50a45d193b6b1c20ed2ae2817261a79a0cc2b5e4bc769707a946" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.590395 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d02b35fa-5c27-4902-99fe-5c7370e34690" path="/var/lib/kubelet/pods/d02b35fa-5c27-4902-99fe-5c7370e34690/volumes" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.607402 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.607730 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfw82\" (UniqueName: \"kubernetes.io/projected/8f59101a-4dfa-4b64-8f24-15041cd24d80-kube-api-access-tfw82\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.607763 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.607950 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-config\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608004 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8f59101a-4dfa-4b64-8f24-15041cd24d80-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608073 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608293 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8f59101a-4dfa-4b64-8f24-15041cd24d80-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608414 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608455 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8f59101a-4dfa-4b64-8f24-15041cd24d80-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608593 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.608734 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8f59101a-4dfa-4b64-8f24-15041cd24d80-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.710453 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.710495 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfw82\" (UniqueName: \"kubernetes.io/projected/8f59101a-4dfa-4b64-8f24-15041cd24d80-kube-api-access-tfw82\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.710519 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.710595 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-config\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711360 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8f59101a-4dfa-4b64-8f24-15041cd24d80-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711424 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711514 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8f59101a-4dfa-4b64-8f24-15041cd24d80-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711591 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711524 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8f59101a-4dfa-4b64-8f24-15041cd24d80-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711624 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8f59101a-4dfa-4b64-8f24-15041cd24d80-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.711858 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8f59101a-4dfa-4b64-8f24-15041cd24d80-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.712098 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.712306 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8f59101a-4dfa-4b64-8f24-15041cd24d80-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.716692 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.717175 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8f59101a-4dfa-4b64-8f24-15041cd24d80-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.717695 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.719376 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-config\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.719404 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.719512 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.719617 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8f59101a-4dfa-4b64-8f24-15041cd24d80-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.720121 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8f59101a-4dfa-4b64-8f24-15041cd24d80-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.730392 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfw82\" (UniqueName: \"kubernetes.io/projected/8f59101a-4dfa-4b64-8f24-15041cd24d80-kube-api-access-tfw82\") pod \"prometheus-metric-storage-0\" (UID: \"8f59101a-4dfa-4b64-8f24-15041cd24d80\") " pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:07 crc kubenswrapper[4796]: I1127 12:38:07.844902 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:08 crc kubenswrapper[4796]: I1127 12:38:08.759238 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 27 12:38:08 crc kubenswrapper[4796]: W1127 12:38:08.777416 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f59101a_4dfa_4b64_8f24_15041cd24d80.slice/crio-423450e52a17a63410e27ec5e2d06a6df9e17f10af1826ccf895b02215efbb29 WatchSource:0}: Error finding container 423450e52a17a63410e27ec5e2d06a6df9e17f10af1826ccf895b02215efbb29: Status 404 returned error can't find the container with id 423450e52a17a63410e27ec5e2d06a6df9e17f10af1826ccf895b02215efbb29 Nov 27 12:38:09 crc kubenswrapper[4796]: I1127 12:38:09.480554 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8f59101a-4dfa-4b64-8f24-15041cd24d80","Type":"ContainerStarted","Data":"423450e52a17a63410e27ec5e2d06a6df9e17f10af1826ccf895b02215efbb29"} Nov 27 12:38:13 crc kubenswrapper[4796]: I1127 12:38:13.516510 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8f59101a-4dfa-4b64-8f24-15041cd24d80","Type":"ContainerStarted","Data":"bac92646112709e12221ce14387013449bff7019c9bfd018448fa192645dd6c7"} Nov 27 12:38:20 crc kubenswrapper[4796]: I1127 12:38:20.580183 4796 generic.go:334] "Generic (PLEG): container finished" podID="8f59101a-4dfa-4b64-8f24-15041cd24d80" containerID="bac92646112709e12221ce14387013449bff7019c9bfd018448fa192645dd6c7" exitCode=0 Nov 27 12:38:20 crc kubenswrapper[4796]: I1127 12:38:20.580303 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8f59101a-4dfa-4b64-8f24-15041cd24d80","Type":"ContainerDied","Data":"bac92646112709e12221ce14387013449bff7019c9bfd018448fa192645dd6c7"} Nov 27 12:38:21 crc kubenswrapper[4796]: I1127 12:38:21.594343 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8f59101a-4dfa-4b64-8f24-15041cd24d80","Type":"ContainerStarted","Data":"a14595c3f37e576f12a281056d85152f710579a4108c8c6c1236b5bfce191fdc"} Nov 27 12:38:23 crc kubenswrapper[4796]: I1127 12:38:23.960192 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m7jgr"] Nov 27 12:38:23 crc kubenswrapper[4796]: I1127 12:38:23.987240 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:23 crc kubenswrapper[4796]: I1127 12:38:23.988594 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m7jgr"] Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.036744 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjlpm\" (UniqueName: \"kubernetes.io/projected/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-kube-api-access-xjlpm\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.036800 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-utilities\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.037028 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-catalog-content\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.138930 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjlpm\" (UniqueName: \"kubernetes.io/projected/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-kube-api-access-xjlpm\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.138994 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-utilities\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.139030 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-catalog-content\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.139701 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-catalog-content\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.139846 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-utilities\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.163646 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjlpm\" (UniqueName: \"kubernetes.io/projected/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-kube-api-access-xjlpm\") pod \"certified-operators-m7jgr\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.312779 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.627044 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8f59101a-4dfa-4b64-8f24-15041cd24d80","Type":"ContainerStarted","Data":"b4cd09b87b3b2943dfa09fc34175570a13dcbbfc817b359667550b017f4d646a"} Nov 27 12:38:24 crc kubenswrapper[4796]: I1127 12:38:24.653539 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m7jgr"] Nov 27 12:38:24 crc kubenswrapper[4796]: W1127 12:38:24.659950 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0f7fee4_c80a_44b7_becf_5877c1ae2c27.slice/crio-347feb964c1df1e0a18fd41dfde15474e4e412edbb09fbd3183f7f1a44163d69 WatchSource:0}: Error finding container 347feb964c1df1e0a18fd41dfde15474e4e412edbb09fbd3183f7f1a44163d69: Status 404 returned error can't find the container with id 347feb964c1df1e0a18fd41dfde15474e4e412edbb09fbd3183f7f1a44163d69 Nov 27 12:38:25 crc kubenswrapper[4796]: I1127 12:38:25.638170 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8f59101a-4dfa-4b64-8f24-15041cd24d80","Type":"ContainerStarted","Data":"e5e3baa12e542c00c34cc8553e5839ce1e339697268bcdbed2d9ac6b1df28d71"} Nov 27 12:38:25 crc kubenswrapper[4796]: I1127 12:38:25.640621 4796 generic.go:334] "Generic (PLEG): container finished" podID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerID="774b9873a0389ad3acc73cd596809352dbf00ac112f460261424451fbc5a182d" exitCode=0 Nov 27 12:38:25 crc kubenswrapper[4796]: I1127 12:38:25.640668 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerDied","Data":"774b9873a0389ad3acc73cd596809352dbf00ac112f460261424451fbc5a182d"} Nov 27 12:38:25 crc kubenswrapper[4796]: I1127 12:38:25.640693 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerStarted","Data":"347feb964c1df1e0a18fd41dfde15474e4e412edbb09fbd3183f7f1a44163d69"} Nov 27 12:38:25 crc kubenswrapper[4796]: I1127 12:38:25.675402 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.675385547 podStartE2EDuration="18.675385547s" podCreationTimestamp="2025-11-27 12:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:38:25.670756331 +0000 UTC m=+4423.189075269" watchObservedRunningTime="2025-11-27 12:38:25.675385547 +0000 UTC m=+4423.193704465" Nov 27 12:38:26 crc kubenswrapper[4796]: I1127 12:38:26.656242 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerStarted","Data":"72512690ce833e2dbe0b5fd8520c6b3d7849068fca8220287765d8c72d5218a6"} Nov 27 12:38:27 crc kubenswrapper[4796]: I1127 12:38:27.669764 4796 generic.go:334] "Generic (PLEG): container finished" podID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerID="72512690ce833e2dbe0b5fd8520c6b3d7849068fca8220287765d8c72d5218a6" exitCode=0 Nov 27 12:38:27 crc kubenswrapper[4796]: I1127 12:38:27.669841 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerDied","Data":"72512690ce833e2dbe0b5fd8520c6b3d7849068fca8220287765d8c72d5218a6"} Nov 27 12:38:27 crc kubenswrapper[4796]: I1127 12:38:27.845832 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:28 crc kubenswrapper[4796]: I1127 12:38:28.685042 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerStarted","Data":"b9d9b08b4b378844bb78f600435b1b10e58f2b68d81e642e353cd54262bcbdd1"} Nov 27 12:38:28 crc kubenswrapper[4796]: I1127 12:38:28.708130 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m7jgr" podStartSLOduration=3.034666821 podStartE2EDuration="5.708109453s" podCreationTimestamp="2025-11-27 12:38:23 +0000 UTC" firstStartedPulling="2025-11-27 12:38:25.643095948 +0000 UTC m=+4423.161414866" lastFinishedPulling="2025-11-27 12:38:28.31653858 +0000 UTC m=+4425.834857498" observedRunningTime="2025-11-27 12:38:28.704677539 +0000 UTC m=+4426.222996467" watchObservedRunningTime="2025-11-27 12:38:28.708109453 +0000 UTC m=+4426.226428391" Nov 27 12:38:34 crc kubenswrapper[4796]: I1127 12:38:34.313899 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:34 crc kubenswrapper[4796]: I1127 12:38:34.314331 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:34 crc kubenswrapper[4796]: I1127 12:38:34.366319 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:34 crc kubenswrapper[4796]: I1127 12:38:34.782675 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.518950 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m7jgr"] Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.520652 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m7jgr" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="registry-server" containerID="cri-o://b9d9b08b4b378844bb78f600435b1b10e58f2b68d81e642e353cd54262bcbdd1" gracePeriod=2 Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.770374 4796 generic.go:334] "Generic (PLEG): container finished" podID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerID="b9d9b08b4b378844bb78f600435b1b10e58f2b68d81e642e353cd54262bcbdd1" exitCode=0 Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.770460 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerDied","Data":"b9d9b08b4b378844bb78f600435b1b10e58f2b68d81e642e353cd54262bcbdd1"} Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.845471 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.854158 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:37 crc kubenswrapper[4796]: I1127 12:38:37.977799 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.127691 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-utilities\") pod \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.128087 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjlpm\" (UniqueName: \"kubernetes.io/projected/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-kube-api-access-xjlpm\") pod \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.128107 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-catalog-content\") pod \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\" (UID: \"d0f7fee4-c80a-44b7-becf-5877c1ae2c27\") " Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.128754 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-utilities" (OuterVolumeSpecName: "utilities") pod "d0f7fee4-c80a-44b7-becf-5877c1ae2c27" (UID: "d0f7fee4-c80a-44b7-becf-5877c1ae2c27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.133379 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-kube-api-access-xjlpm" (OuterVolumeSpecName: "kube-api-access-xjlpm") pod "d0f7fee4-c80a-44b7-becf-5877c1ae2c27" (UID: "d0f7fee4-c80a-44b7-becf-5877c1ae2c27"). InnerVolumeSpecName "kube-api-access-xjlpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.174778 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0f7fee4-c80a-44b7-becf-5877c1ae2c27" (UID: "d0f7fee4-c80a-44b7-becf-5877c1ae2c27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.230937 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjlpm\" (UniqueName: \"kubernetes.io/projected/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-kube-api-access-xjlpm\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.231654 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.231822 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0f7fee4-c80a-44b7-becf-5877c1ae2c27-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.785775 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7jgr" event={"ID":"d0f7fee4-c80a-44b7-becf-5877c1ae2c27","Type":"ContainerDied","Data":"347feb964c1df1e0a18fd41dfde15474e4e412edbb09fbd3183f7f1a44163d69"} Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.785826 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m7jgr" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.785831 4796 scope.go:117] "RemoveContainer" containerID="b9d9b08b4b378844bb78f600435b1b10e58f2b68d81e642e353cd54262bcbdd1" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.792937 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.821455 4796 scope.go:117] "RemoveContainer" containerID="72512690ce833e2dbe0b5fd8520c6b3d7849068fca8220287765d8c72d5218a6" Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.862587 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m7jgr"] Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.870602 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m7jgr"] Nov 27 12:38:38 crc kubenswrapper[4796]: I1127 12:38:38.992368 4796 scope.go:117] "RemoveContainer" containerID="774b9873a0389ad3acc73cd596809352dbf00ac112f460261424451fbc5a182d" Nov 27 12:38:39 crc kubenswrapper[4796]: I1127 12:38:39.580285 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" path="/var/lib/kubelet/pods/d0f7fee4-c80a-44b7-becf-5877c1ae2c27/volumes" Nov 27 12:40:05 crc kubenswrapper[4796]: I1127 12:40:05.313944 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.298711 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-76h9f/must-gather-62n49"] Nov 27 12:40:24 crc kubenswrapper[4796]: E1127 12:40:24.299713 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="extract-utilities" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.299734 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="extract-utilities" Nov 27 12:40:24 crc kubenswrapper[4796]: E1127 12:40:24.299761 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="registry-server" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.299775 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="registry-server" Nov 27 12:40:24 crc kubenswrapper[4796]: E1127 12:40:24.299785 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="extract-content" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.299791 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="extract-content" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.299985 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f7fee4-c80a-44b7-becf-5877c1ae2c27" containerName="registry-server" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.301352 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.303200 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-76h9f"/"default-dockercfg-cx7tn" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.304244 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-76h9f"/"kube-root-ca.crt" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.304455 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-76h9f"/"openshift-service-ca.crt" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.333299 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-76h9f/must-gather-62n49"] Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.392053 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2t47\" (UniqueName: \"kubernetes.io/projected/907cb4cc-62e8-45fa-88a1-5226e75b11d2-kube-api-access-q2t47\") pod \"must-gather-62n49\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.392126 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/907cb4cc-62e8-45fa-88a1-5226e75b11d2-must-gather-output\") pod \"must-gather-62n49\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.493542 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2t47\" (UniqueName: \"kubernetes.io/projected/907cb4cc-62e8-45fa-88a1-5226e75b11d2-kube-api-access-q2t47\") pod \"must-gather-62n49\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.493594 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/907cb4cc-62e8-45fa-88a1-5226e75b11d2-must-gather-output\") pod \"must-gather-62n49\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.494090 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/907cb4cc-62e8-45fa-88a1-5226e75b11d2-must-gather-output\") pod \"must-gather-62n49\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.516980 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2t47\" (UniqueName: \"kubernetes.io/projected/907cb4cc-62e8-45fa-88a1-5226e75b11d2-kube-api-access-q2t47\") pod \"must-gather-62n49\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:24 crc kubenswrapper[4796]: I1127 12:40:24.631018 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:40:25 crc kubenswrapper[4796]: I1127 12:40:25.072774 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-76h9f/must-gather-62n49"] Nov 27 12:40:25 crc kubenswrapper[4796]: I1127 12:40:25.928522 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/must-gather-62n49" event={"ID":"907cb4cc-62e8-45fa-88a1-5226e75b11d2","Type":"ContainerStarted","Data":"7faca9529139d436589e8d9090e30860e9711923e69912a701a8baf2f98e0ded"} Nov 27 12:40:31 crc kubenswrapper[4796]: I1127 12:40:31.880310 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:40:31 crc kubenswrapper[4796]: I1127 12:40:31.880904 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:40:33 crc kubenswrapper[4796]: I1127 12:40:33.014546 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/must-gather-62n49" event={"ID":"907cb4cc-62e8-45fa-88a1-5226e75b11d2","Type":"ContainerStarted","Data":"1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4"} Nov 27 12:40:33 crc kubenswrapper[4796]: I1127 12:40:33.015020 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/must-gather-62n49" event={"ID":"907cb4cc-62e8-45fa-88a1-5226e75b11d2","Type":"ContainerStarted","Data":"0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70"} Nov 27 12:40:33 crc kubenswrapper[4796]: I1127 12:40:33.029556 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-76h9f/must-gather-62n49" podStartSLOduration=1.941101616 podStartE2EDuration="9.029538307s" podCreationTimestamp="2025-11-27 12:40:24 +0000 UTC" firstStartedPulling="2025-11-27 12:40:25.079146466 +0000 UTC m=+4542.597465384" lastFinishedPulling="2025-11-27 12:40:32.167583157 +0000 UTC m=+4549.685902075" observedRunningTime="2025-11-27 12:40:33.026012912 +0000 UTC m=+4550.544331830" watchObservedRunningTime="2025-11-27 12:40:33.029538307 +0000 UTC m=+4550.547857225" Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.519005 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-76h9f/crc-debug-5h6z2"] Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.520661 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.570755 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1ca74175-81f1-4368-a61c-a57344f421e3-host\") pod \"crc-debug-5h6z2\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.570798 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz5cq\" (UniqueName: \"kubernetes.io/projected/1ca74175-81f1-4368-a61c-a57344f421e3-kube-api-access-zz5cq\") pod \"crc-debug-5h6z2\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.672125 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1ca74175-81f1-4368-a61c-a57344f421e3-host\") pod \"crc-debug-5h6z2\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.672697 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1ca74175-81f1-4368-a61c-a57344f421e3-host\") pod \"crc-debug-5h6z2\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:39 crc kubenswrapper[4796]: I1127 12:40:39.672966 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz5cq\" (UniqueName: \"kubernetes.io/projected/1ca74175-81f1-4368-a61c-a57344f421e3-kube-api-access-zz5cq\") pod \"crc-debug-5h6z2\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:40 crc kubenswrapper[4796]: I1127 12:40:40.073921 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz5cq\" (UniqueName: \"kubernetes.io/projected/1ca74175-81f1-4368-a61c-a57344f421e3-kube-api-access-zz5cq\") pod \"crc-debug-5h6z2\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:40 crc kubenswrapper[4796]: I1127 12:40:40.144949 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:40:41 crc kubenswrapper[4796]: I1127 12:40:41.101791 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" event={"ID":"1ca74175-81f1-4368-a61c-a57344f421e3","Type":"ContainerStarted","Data":"7427be33142f3cf25783c51f50b4cbe71154501641bd8038c0923e15a72ebf73"} Nov 27 12:40:57 crc kubenswrapper[4796]: I1127 12:40:57.308952 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" event={"ID":"1ca74175-81f1-4368-a61c-a57344f421e3","Type":"ContainerStarted","Data":"bbf2fe8905fcfa81f86809192981816b48a96bbe29771b1d25488c1ea2050c4e"} Nov 27 12:40:57 crc kubenswrapper[4796]: I1127 12:40:57.333240 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" podStartSLOduration=2.09835266 podStartE2EDuration="18.333217632s" podCreationTimestamp="2025-11-27 12:40:39 +0000 UTC" firstStartedPulling="2025-11-27 12:40:40.200781051 +0000 UTC m=+4557.719099969" lastFinishedPulling="2025-11-27 12:40:56.435646023 +0000 UTC m=+4573.953964941" observedRunningTime="2025-11-27 12:40:57.323185889 +0000 UTC m=+4574.841504807" watchObservedRunningTime="2025-11-27 12:40:57.333217632 +0000 UTC m=+4574.851536570" Nov 27 12:41:01 crc kubenswrapper[4796]: I1127 12:41:01.879944 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:41:01 crc kubenswrapper[4796]: I1127 12:41:01.880480 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:41:18 crc kubenswrapper[4796]: I1127 12:41:18.508966 4796 generic.go:334] "Generic (PLEG): container finished" podID="1ca74175-81f1-4368-a61c-a57344f421e3" containerID="bbf2fe8905fcfa81f86809192981816b48a96bbe29771b1d25488c1ea2050c4e" exitCode=0 Nov 27 12:41:18 crc kubenswrapper[4796]: I1127 12:41:18.509045 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" event={"ID":"1ca74175-81f1-4368-a61c-a57344f421e3","Type":"ContainerDied","Data":"bbf2fe8905fcfa81f86809192981816b48a96bbe29771b1d25488c1ea2050c4e"} Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.654931 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.713517 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-76h9f/crc-debug-5h6z2"] Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.721904 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-76h9f/crc-debug-5h6z2"] Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.793758 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz5cq\" (UniqueName: \"kubernetes.io/projected/1ca74175-81f1-4368-a61c-a57344f421e3-kube-api-access-zz5cq\") pod \"1ca74175-81f1-4368-a61c-a57344f421e3\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.793899 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1ca74175-81f1-4368-a61c-a57344f421e3-host\") pod \"1ca74175-81f1-4368-a61c-a57344f421e3\" (UID: \"1ca74175-81f1-4368-a61c-a57344f421e3\") " Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.794290 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca74175-81f1-4368-a61c-a57344f421e3-host" (OuterVolumeSpecName: "host") pod "1ca74175-81f1-4368-a61c-a57344f421e3" (UID: "1ca74175-81f1-4368-a61c-a57344f421e3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.794636 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1ca74175-81f1-4368-a61c-a57344f421e3-host\") on node \"crc\" DevicePath \"\"" Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.798733 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca74175-81f1-4368-a61c-a57344f421e3-kube-api-access-zz5cq" (OuterVolumeSpecName: "kube-api-access-zz5cq") pod "1ca74175-81f1-4368-a61c-a57344f421e3" (UID: "1ca74175-81f1-4368-a61c-a57344f421e3"). InnerVolumeSpecName "kube-api-access-zz5cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:41:19 crc kubenswrapper[4796]: I1127 12:41:19.896470 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz5cq\" (UniqueName: \"kubernetes.io/projected/1ca74175-81f1-4368-a61c-a57344f421e3-kube-api-access-zz5cq\") on node \"crc\" DevicePath \"\"" Nov 27 12:41:20 crc kubenswrapper[4796]: I1127 12:41:20.546170 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7427be33142f3cf25783c51f50b4cbe71154501641bd8038c0923e15a72ebf73" Nov 27 12:41:20 crc kubenswrapper[4796]: I1127 12:41:20.546248 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-5h6z2" Nov 27 12:41:20 crc kubenswrapper[4796]: E1127 12:41:20.682292 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ca74175_81f1_4368_a61c_a57344f421e3.slice/crio-7427be33142f3cf25783c51f50b4cbe71154501641bd8038c0923e15a72ebf73\": RecentStats: unable to find data in memory cache]" Nov 27 12:41:20 crc kubenswrapper[4796]: I1127 12:41:20.916774 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-76h9f/crc-debug-dqdqj"] Nov 27 12:41:20 crc kubenswrapper[4796]: E1127 12:41:20.917186 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca74175-81f1-4368-a61c-a57344f421e3" containerName="container-00" Nov 27 12:41:20 crc kubenswrapper[4796]: I1127 12:41:20.917203 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca74175-81f1-4368-a61c-a57344f421e3" containerName="container-00" Nov 27 12:41:20 crc kubenswrapper[4796]: I1127 12:41:20.917647 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca74175-81f1-4368-a61c-a57344f421e3" containerName="container-00" Nov 27 12:41:20 crc kubenswrapper[4796]: I1127 12:41:20.918367 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.017227 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zhjz\" (UniqueName: \"kubernetes.io/projected/9e2fc84c-1468-4627-b72b-47f9b2f5d630-kube-api-access-6zhjz\") pod \"crc-debug-dqdqj\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.017281 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e2fc84c-1468-4627-b72b-47f9b2f5d630-host\") pod \"crc-debug-dqdqj\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.119151 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zhjz\" (UniqueName: \"kubernetes.io/projected/9e2fc84c-1468-4627-b72b-47f9b2f5d630-kube-api-access-6zhjz\") pod \"crc-debug-dqdqj\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.119196 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e2fc84c-1468-4627-b72b-47f9b2f5d630-host\") pod \"crc-debug-dqdqj\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.119306 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e2fc84c-1468-4627-b72b-47f9b2f5d630-host\") pod \"crc-debug-dqdqj\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.139964 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zhjz\" (UniqueName: \"kubernetes.io/projected/9e2fc84c-1468-4627-b72b-47f9b2f5d630-kube-api-access-6zhjz\") pod \"crc-debug-dqdqj\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.242161 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:21 crc kubenswrapper[4796]: W1127 12:41:21.276720 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e2fc84c_1468_4627_b72b_47f9b2f5d630.slice/crio-c995dc4618aec699603d1072e90840d947fe98d64027d57dbaf13ad8b0d92c9c WatchSource:0}: Error finding container c995dc4618aec699603d1072e90840d947fe98d64027d57dbaf13ad8b0d92c9c: Status 404 returned error can't find the container with id c995dc4618aec699603d1072e90840d947fe98d64027d57dbaf13ad8b0d92c9c Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.562576 4796 generic.go:334] "Generic (PLEG): container finished" podID="9e2fc84c-1468-4627-b72b-47f9b2f5d630" containerID="d53ac62f8e7360b57447a2618fd5ffd32cdde59a3f8959ab2dce3547a6d89c28" exitCode=1 Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.562649 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/crc-debug-dqdqj" event={"ID":"9e2fc84c-1468-4627-b72b-47f9b2f5d630","Type":"ContainerDied","Data":"d53ac62f8e7360b57447a2618fd5ffd32cdde59a3f8959ab2dce3547a6d89c28"} Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.563294 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/crc-debug-dqdqj" event={"ID":"9e2fc84c-1468-4627-b72b-47f9b2f5d630","Type":"ContainerStarted","Data":"c995dc4618aec699603d1072e90840d947fe98d64027d57dbaf13ad8b0d92c9c"} Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.584005 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca74175-81f1-4368-a61c-a57344f421e3" path="/var/lib/kubelet/pods/1ca74175-81f1-4368-a61c-a57344f421e3/volumes" Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.599953 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-76h9f/crc-debug-dqdqj"] Nov 27 12:41:21 crc kubenswrapper[4796]: I1127 12:41:21.608403 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-76h9f/crc-debug-dqdqj"] Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.676655 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.750066 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zhjz\" (UniqueName: \"kubernetes.io/projected/9e2fc84c-1468-4627-b72b-47f9b2f5d630-kube-api-access-6zhjz\") pod \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.750224 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e2fc84c-1468-4627-b72b-47f9b2f5d630-host\") pod \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\" (UID: \"9e2fc84c-1468-4627-b72b-47f9b2f5d630\") " Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.750631 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e2fc84c-1468-4627-b72b-47f9b2f5d630-host" (OuterVolumeSpecName: "host") pod "9e2fc84c-1468-4627-b72b-47f9b2f5d630" (UID: "9e2fc84c-1468-4627-b72b-47f9b2f5d630"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.751024 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e2fc84c-1468-4627-b72b-47f9b2f5d630-host\") on node \"crc\" DevicePath \"\"" Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.757842 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2fc84c-1468-4627-b72b-47f9b2f5d630-kube-api-access-6zhjz" (OuterVolumeSpecName: "kube-api-access-6zhjz") pod "9e2fc84c-1468-4627-b72b-47f9b2f5d630" (UID: "9e2fc84c-1468-4627-b72b-47f9b2f5d630"). InnerVolumeSpecName "kube-api-access-6zhjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:41:22 crc kubenswrapper[4796]: I1127 12:41:22.853195 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zhjz\" (UniqueName: \"kubernetes.io/projected/9e2fc84c-1468-4627-b72b-47f9b2f5d630-kube-api-access-6zhjz\") on node \"crc\" DevicePath \"\"" Nov 27 12:41:23 crc kubenswrapper[4796]: I1127 12:41:23.579148 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e2fc84c-1468-4627-b72b-47f9b2f5d630" path="/var/lib/kubelet/pods/9e2fc84c-1468-4627-b72b-47f9b2f5d630/volumes" Nov 27 12:41:23 crc kubenswrapper[4796]: I1127 12:41:23.581413 4796 scope.go:117] "RemoveContainer" containerID="d53ac62f8e7360b57447a2618fd5ffd32cdde59a3f8959ab2dce3547a6d89c28" Nov 27 12:41:23 crc kubenswrapper[4796]: I1127 12:41:23.581548 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/crc-debug-dqdqj" Nov 27 12:41:31 crc kubenswrapper[4796]: I1127 12:41:31.879546 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:41:31 crc kubenswrapper[4796]: I1127 12:41:31.880175 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:41:31 crc kubenswrapper[4796]: I1127 12:41:31.880236 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:41:31 crc kubenswrapper[4796]: I1127 12:41:31.881242 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:41:31 crc kubenswrapper[4796]: I1127 12:41:31.881346 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" gracePeriod=600 Nov 27 12:41:32 crc kubenswrapper[4796]: E1127 12:41:32.014250 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:41:32 crc kubenswrapper[4796]: I1127 12:41:32.699192 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" exitCode=0 Nov 27 12:41:32 crc kubenswrapper[4796]: I1127 12:41:32.699233 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648"} Nov 27 12:41:32 crc kubenswrapper[4796]: I1127 12:41:32.699266 4796 scope.go:117] "RemoveContainer" containerID="2c9f964331a818e21cdd07be85bc6c8a01ad0977a6954ef3cab78c4971adbb4e" Nov 27 12:41:32 crc kubenswrapper[4796]: I1127 12:41:32.700048 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:41:32 crc kubenswrapper[4796]: E1127 12:41:32.700388 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:41:41 crc kubenswrapper[4796]: E1127 12:41:41.233663 4796 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 27 12:41:47 crc kubenswrapper[4796]: I1127 12:41:47.569251 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:41:47 crc kubenswrapper[4796]: E1127 12:41:47.570123 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:41:59 crc kubenswrapper[4796]: I1127 12:41:59.574363 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:41:59 crc kubenswrapper[4796]: E1127 12:41:59.575110 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:42:07 crc kubenswrapper[4796]: I1127 12:42:07.626668 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/init-config-reloader/0.log" Nov 27 12:42:07 crc kubenswrapper[4796]: I1127 12:42:07.816647 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/init-config-reloader/0.log" Nov 27 12:42:07 crc kubenswrapper[4796]: I1127 12:42:07.835502 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/config-reloader/0.log" Nov 27 12:42:07 crc kubenswrapper[4796]: I1127 12:42:07.847594 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/alertmanager/0.log" Nov 27 12:42:08 crc kubenswrapper[4796]: I1127 12:42:08.055354 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-api/0.log" Nov 27 12:42:08 crc kubenswrapper[4796]: I1127 12:42:08.056864 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-evaluator/0.log" Nov 27 12:42:08 crc kubenswrapper[4796]: I1127 12:42:08.077053 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-listener/0.log" Nov 27 12:42:08 crc kubenswrapper[4796]: I1127 12:42:08.457881 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-notifier/0.log" Nov 27 12:42:08 crc kubenswrapper[4796]: I1127 12:42:08.483999 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-655d9778b6-79p2x_c71cc6f0-df01-4c3d-9962-29d71307f38a/barbican-api/0.log" Nov 27 12:42:08 crc kubenswrapper[4796]: I1127 12:42:08.518672 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-655d9778b6-79p2x_c71cc6f0-df01-4c3d-9962-29d71307f38a/barbican-api-log/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.295522 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79c6986754-nzz2p_c0327df8-c909-4f47-bff7-7519ded4ea2a/barbican-keystone-listener/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.355104 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79c6986754-nzz2p_c0327df8-c909-4f47-bff7-7519ded4ea2a/barbican-keystone-listener-log/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.521774 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7968f8455-ntchl_176ad434-447c-4540-ac74-d80d1aeab2fd/barbican-worker/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.541645 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7968f8455-ntchl_176ad434-447c-4540-ac74-d80d1aeab2fd/barbican-worker-log/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.705642 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx_05691e2e-3935-4eef-9d8b-f591516ebb3c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.785025 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/ceilometer-central-agent/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.837138 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/ceilometer-notification-agent/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.912292 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/proxy-httpd/0.log" Nov 27 12:42:09 crc kubenswrapper[4796]: I1127 12:42:09.995174 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/sg-core/0.log" Nov 27 12:42:10 crc kubenswrapper[4796]: I1127 12:42:10.138231 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7ae7bbf2-68f9-489d-ab84-37202429a5fb/cinder-api/0.log" Nov 27 12:42:10 crc kubenswrapper[4796]: I1127 12:42:10.185307 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7ae7bbf2-68f9-489d-ab84-37202429a5fb/cinder-api-log/0.log" Nov 27 12:42:10 crc kubenswrapper[4796]: I1127 12:42:10.353335 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f8373371-1f42-4460-8724-668392be8975/cinder-scheduler/0.log" Nov 27 12:42:10 crc kubenswrapper[4796]: I1127 12:42:10.365969 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f8373371-1f42-4460-8724-668392be8975/probe/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.002697 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-krh4d_cb67680b-c997-4002-a238-67ff005b15f0/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.007159 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kfn44_c41ab70c-1f9c-4961-9e0c-eea2a846f125/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.126541 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-kdrpj_7dfd0463-f427-4277-b02d-444ef9bd8565/init/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.410595 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-kdrpj_7dfd0463-f427-4277-b02d-444ef9bd8565/dnsmasq-dns/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.414586 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-kdrpj_7dfd0463-f427-4277-b02d-444ef9bd8565/init/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.454571 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm_72e23875-5131-453e-a9d5-dcdbea0b4cf9/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.626378 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1f85834f-7915-4c31-b55d-5c4948763a3c/glance-log/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.640817 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1f85834f-7915-4c31-b55d-5c4948763a3c/glance-httpd/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.716336 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f7681c6a-f516-4b70-848c-2a6ce2c2e087/glance-httpd/0.log" Nov 27 12:42:11 crc kubenswrapper[4796]: I1127 12:42:11.735368 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f7681c6a-f516-4b70-848c-2a6ce2c2e087/glance-log/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.106729 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5469764df4-8wt46_ecd41927-ca66-42f8-871b-2b872ba9c3c8/heat-engine/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.277203 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6896844f76-vnbq9_acf45387-6183-41dd-848c-160078149bfd/heat-api/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.359919 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5b7ccbdf6f-g754q_6e828f09-14e9-444f-9ef6-6a07f5f665e5/heat-cfnapi/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.452300 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb_77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.569152 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:42:12 crc kubenswrapper[4796]: E1127 12:42:12.569506 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.591614 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bj5x7_62906f9c-c8ce-48d2-b38b-db707e82f24c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.717314 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29404081-jlkpb_48aeb330-d74c-4814-9c4d-41e40436e8a3/keystone-cron/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.774626 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-95c9f9f9b-zshfm_0e86b961-a118-4ef7-a58a-a1a350c48783/keystone-api/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.866889 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_82fecd54-d3de-4157-b04e-3187be01c0ad/kube-state-metrics/0.log" Nov 27 12:42:12 crc kubenswrapper[4796]: I1127 12:42:12.968981 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8_a779ff89-a04b-41a9-acfe-64e318bbf745/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:13 crc kubenswrapper[4796]: I1127 12:42:13.278023 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69bc59f7f-g45f2_cdd9f25b-34db-477b-a6c6-11a5732cacdc/neutron-api/0.log" Nov 27 12:42:13 crc kubenswrapper[4796]: I1127 12:42:13.312843 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69bc59f7f-g45f2_cdd9f25b-34db-477b-a6c6-11a5732cacdc/neutron-httpd/0.log" Nov 27 12:42:13 crc kubenswrapper[4796]: I1127 12:42:13.427390 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z_213166c4-bd90-4e0f-8fbb-1164ebc31cb8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:13 crc kubenswrapper[4796]: I1127 12:42:13.848625 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d4574506-0fa9-41e7-ace2-0151a43916a5/nova-api-log/0.log" Nov 27 12:42:13 crc kubenswrapper[4796]: I1127 12:42:13.955588 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a30067ee-e2bf-4d5f-bdcf-cff461743b62/nova-cell0-conductor-conductor/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.026931 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d4574506-0fa9-41e7-ace2-0151a43916a5/nova-api-api/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.167439 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_97d375ad-8df3-4053-a31a-1483cbc5ce30/nova-cell1-conductor-conductor/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.309686 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_69f6622f-9286-49ea-b8b7-f56a61143e03/nova-cell1-novncproxy-novncproxy/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.530908 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-42r5d_bc31c0f1-6642-4ff6-8fca-16c7ca6b4941/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.638461 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_12250a93-d356-4097-b50e-57343bc8b2c9/nova-metadata-log/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.880706 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_b47781b1-daeb-4448-b7c5-55005abc5f2e/nova-scheduler-scheduler/0.log" Nov 27 12:42:14 crc kubenswrapper[4796]: I1127 12:42:14.978160 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bcc2f71a-d686-4c12-adc3-6a18ca6225f8/mysql-bootstrap/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.197648 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bcc2f71a-d686-4c12-adc3-6a18ca6225f8/mysql-bootstrap/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.198154 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bcc2f71a-d686-4c12-adc3-6a18ca6225f8/galera/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.412695 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8f3e487f-6783-499b-bd57-0a811f4ef5f8/mysql-bootstrap/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.642005 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8f3e487f-6783-499b-bd57-0a811f4ef5f8/galera/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.654204 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8f3e487f-6783-499b-bd57-0a811f4ef5f8/mysql-bootstrap/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.778180 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_12250a93-d356-4097-b50e-57343bc8b2c9/nova-metadata-metadata/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.804966 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4bdf142a-761b-4879-b80b-b23c3f974593/openstackclient/0.log" Nov 27 12:42:15 crc kubenswrapper[4796]: I1127 12:42:15.875513 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-q2vvb_d5cd65c9-3282-4039-9963-f1174a310831/openstack-network-exporter/0.log" Nov 27 12:42:16 crc kubenswrapper[4796]: I1127 12:42:16.250687 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovsdb-server-init/0.log" Nov 27 12:42:16 crc kubenswrapper[4796]: I1127 12:42:16.624884 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovsdb-server-init/0.log" Nov 27 12:42:16 crc kubenswrapper[4796]: I1127 12:42:16.649259 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovsdb-server/0.log" Nov 27 12:42:16 crc kubenswrapper[4796]: I1127 12:42:16.717138 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovs-vswitchd/0.log" Nov 27 12:42:16 crc kubenswrapper[4796]: I1127 12:42:16.899498 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zjmwj_99ef6c07-3c37-4a97-8fbc-bd09a91617a0/ovn-controller/0.log" Nov 27 12:42:16 crc kubenswrapper[4796]: I1127 12:42:16.983942 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7gg7l_4b5e628e-9f89-4ae6-80c0-74d9cf57535d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.210349 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_469ea57e-e812-49e4-a3f3-947bf53deb22/openstack-network-exporter/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.265616 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_469ea57e-e812-49e4-a3f3-947bf53deb22/ovn-northd/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.285109 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8999de5c-3782-44dc-b88d-29f80c31a55f/openstack-network-exporter/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.453300 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8999de5c-3782-44dc-b88d-29f80c31a55f/ovsdbserver-nb/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.479049 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_19c0f33c-82a7-4987-be98-e699c9d32051/openstack-network-exporter/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.518553 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_19c0f33c-82a7-4987-be98-e699c9d32051/ovsdbserver-sb/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.916443 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d4b54c868-49fn9_c9233c70-4c77-41e8-831b-682307ae7dd7/placement-api/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.936688 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d4b54c868-49fn9_c9233c70-4c77-41e8-831b-682307ae7dd7/placement-log/0.log" Nov 27 12:42:17 crc kubenswrapper[4796]: I1127 12:42:17.952663 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/init-config-reloader/0.log" Nov 27 12:42:18 crc kubenswrapper[4796]: I1127 12:42:18.578402 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/init-config-reloader/0.log" Nov 27 12:42:18 crc kubenswrapper[4796]: I1127 12:42:18.615142 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/config-reloader/0.log" Nov 27 12:42:18 crc kubenswrapper[4796]: I1127 12:42:18.644172 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/thanos-sidecar/0.log" Nov 27 12:42:18 crc kubenswrapper[4796]: I1127 12:42:18.645594 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/prometheus/0.log" Nov 27 12:42:18 crc kubenswrapper[4796]: I1127 12:42:18.827353 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4f2a8e-1ba5-4af3-8154-02a911ca9d14/setup-container/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.085627 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4f2a8e-1ba5-4af3-8154-02a911ca9d14/rabbitmq/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.172366 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a340b9d9-d859-4b22-a55a-e955b7951806/setup-container/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.220611 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4f2a8e-1ba5-4af3-8154-02a911ca9d14/setup-container/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.439638 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a340b9d9-d859-4b22-a55a-e955b7951806/setup-container/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.457832 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a340b9d9-d859-4b22-a55a-e955b7951806/rabbitmq/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.470403 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw_bf61d6a5-49e7-4dc9-be44-0292579185c1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.718360 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb_81f86614-5247-4cb9-ad8c-688c70f862a6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:19 crc kubenswrapper[4796]: I1127 12:42:19.726445 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s45b6_9059a6c5-d130-4751-849d-dc05d1c5c037/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:20 crc kubenswrapper[4796]: I1127 12:42:20.641821 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-ldbkm_eb6f1515-67d6-4070-ad15-fb26e89835d8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:20 crc kubenswrapper[4796]: I1127 12:42:20.687846 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ltbbg_6b7f9b64-f1d9-4468-aafd-30819c86ab8a/ssh-known-hosts-edpm-deployment/0.log" Nov 27 12:42:20 crc kubenswrapper[4796]: I1127 12:42:20.974345 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6568564dcc-mst5g_c65b890a-db52-4342-938c-d56538e76b3c/proxy-server/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.056604 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6568564dcc-mst5g_c65b890a-db52-4342-938c-d56538e76b3c/proxy-httpd/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.228811 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jw9v8_085a80b3-9f90-480a-8e26-fda7fb500c64/swift-ring-rebalance/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.325967 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-reaper/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.346131 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-auditor/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.456552 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-server/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.482197 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-replicator/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.614060 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-auditor/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.629870 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-replicator/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.718484 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-server/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.789783 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-updater/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.899064 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-expirer/0.log" Nov 27 12:42:21 crc kubenswrapper[4796]: I1127 12:42:21.954455 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-auditor/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.013772 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-replicator/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.054525 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-server/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.123578 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-updater/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.153772 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/rsync/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.222520 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/swift-recon-cron/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.403946 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k_72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:22 crc kubenswrapper[4796]: I1127 12:42:22.502071 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m_8df61624-3a9b-4f6c-aa8a-2843e80fe64d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:42:26 crc kubenswrapper[4796]: I1127 12:42:26.569231 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:42:26 crc kubenswrapper[4796]: E1127 12:42:26.570032 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:42:29 crc kubenswrapper[4796]: I1127 12:42:29.967532 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21/memcached/0.log" Nov 27 12:42:41 crc kubenswrapper[4796]: I1127 12:42:41.568883 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:42:41 crc kubenswrapper[4796]: E1127 12:42:41.569826 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.114442 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/util/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.343205 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/util/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.351898 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/pull/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.393967 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/pull/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.538304 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/extract/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.545937 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/util/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.560581 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/pull/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.759701 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-z5fwq_ca4033f3-d225-4700-bf24-875462a1a404/manager/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.769452 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-z5fwq_ca4033f3-d225-4700-bf24-875462a1a404/kube-rbac-proxy/0.log" Nov 27 12:42:51 crc kubenswrapper[4796]: I1127 12:42:51.803610 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-lcrbs_154fc49c-3eca-4d26-836f-ed87d2d78716/kube-rbac-proxy/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.092057 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-78jwg_efe899c4-0891-480f-9aaa-f8b795706142/kube-rbac-proxy/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.163376 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-78jwg_efe899c4-0891-480f-9aaa-f8b795706142/manager/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.181322 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-lcrbs_154fc49c-3eca-4d26-836f-ed87d2d78716/manager/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.284654 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-zlbmz_0c41793c-bff2-4a6a-877c-489df83fa578/kube-rbac-proxy/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.471565 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-zlbmz_0c41793c-bff2-4a6a-877c-489df83fa578/manager/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.527642 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-zqg7w_2b095ec7-98b1-4e3c-89ed-0bec53da6057/kube-rbac-proxy/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.569810 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-zqg7w_2b095ec7-98b1-4e3c-89ed-0bec53da6057/manager/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.673843 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-wkg8j_24bc27f0-611d-4147-9819-6ec0eb012d81/kube-rbac-proxy/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.703338 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-wkg8j_24bc27f0-611d-4147-9819-6ec0eb012d81/manager/0.log" Nov 27 12:42:52 crc kubenswrapper[4796]: I1127 12:42:52.844774 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zzhd9_940eec7d-4d26-462b-958b-ec782cf28fdf/kube-rbac-proxy/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.000243 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-6fpn9_0b1822e4-eff9-4ae6-9d91-46bbef3d090b/kube-rbac-proxy/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.058028 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-6fpn9_0b1822e4-eff9-4ae6-9d91-46bbef3d090b/manager/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.063330 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zzhd9_940eec7d-4d26-462b-958b-ec782cf28fdf/manager/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.795201 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-drlr5_5865e45d-8c50-45ea-9fb5-f8a14385de1a/kube-rbac-proxy/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.855481 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-drlr5_5865e45d-8c50-45ea-9fb5-f8a14385de1a/manager/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.963209 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-ll522_6858cdee-21cc-4595-905a-3cd4ce0da27c/kube-rbac-proxy/0.log" Nov 27 12:42:53 crc kubenswrapper[4796]: I1127 12:42:53.963952 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-ll522_6858cdee-21cc-4595-905a-3cd4ce0da27c/manager/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.060196 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-8wjn5_257e6c0f-c3fd-458e-839a-d28ca580fd13/kube-rbac-proxy/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.161355 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-8wjn5_257e6c0f-c3fd-458e-839a-d28ca580fd13/manager/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.221056 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-fsmqp_d8beead6-b3d4-4f4c-baae-4eaf00fe812a/kube-rbac-proxy/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.314357 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-fsmqp_d8beead6-b3d4-4f4c-baae-4eaf00fe812a/manager/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.377518 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6ks4r_2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1/kube-rbac-proxy/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.493841 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6ks4r_2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1/manager/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.552366 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-8lq5v_25fd5653-6347-40b2-9d4e-531b58a7e6a3/kube-rbac-proxy/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.603418 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-8lq5v_25fd5653-6347-40b2-9d4e-531b58a7e6a3/manager/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.731072 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w_c2e58505-5b12-41ac-aa51-d857b46b0166/kube-rbac-proxy/0.log" Nov 27 12:42:54 crc kubenswrapper[4796]: I1127 12:42:54.749968 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w_c2e58505-5b12-41ac-aa51-d857b46b0166/manager/0.log" Nov 27 12:42:55 crc kubenswrapper[4796]: I1127 12:42:55.569809 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:42:55 crc kubenswrapper[4796]: E1127 12:42:55.571061 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:42:55 crc kubenswrapper[4796]: I1127 12:42:55.632392 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rbwz6_037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5/registry-server/0.log" Nov 27 12:42:55 crc kubenswrapper[4796]: I1127 12:42:55.706675 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68f75d5585-g6s7q_b8c1db68-d88e-406c-a551-bfb6fe0df332/operator/0.log" Nov 27 12:42:55 crc kubenswrapper[4796]: I1127 12:42:55.918167 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-4xs4w_b0ba2f52-f61b-47b5-8739-3c248d6e012b/manager/0.log" Nov 27 12:42:55 crc kubenswrapper[4796]: I1127 12:42:55.923869 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-4xs4w_b0ba2f52-f61b-47b5-8739-3c248d6e012b/kube-rbac-proxy/0.log" Nov 27 12:42:55 crc kubenswrapper[4796]: I1127 12:42:55.924333 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-cjtbm_3189b882-80fe-4198-872d-b8b36f0cc766/kube-rbac-proxy/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.171834 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-l2v4r_6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb/operator/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.200320 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-cjtbm_3189b882-80fe-4198-872d-b8b36f0cc766/manager/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.400613 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-grznn_b72bebe8-8cf0-4e6e-908b-a5095a7a9348/kube-rbac-proxy/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.419939 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-grznn_b72bebe8-8cf0-4e6e-908b-a5095a7a9348/manager/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.438518 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557965c5b6-9qjgg_12be3ae6-8893-4998-94bd-fd12c7c6640c/manager/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.508473 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/kube-rbac-proxy/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.632075 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-d8k66_240033a0-6f60-4c73-b2cb-8d4ddb91a66d/kube-rbac-proxy/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.684507 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-d8k66_240033a0-6f60-4c73-b2cb-8d4ddb91a66d/manager/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.743332 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.810119 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-88bxr_4905ae06-8f67-4beb-93ca-bcd9e1dca63e/kube-rbac-proxy/0.log" Nov 27 12:42:56 crc kubenswrapper[4796]: I1127 12:42:56.836008 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-88bxr_4905ae06-8f67-4beb-93ca-bcd9e1dca63e/manager/0.log" Nov 27 12:43:09 crc kubenswrapper[4796]: I1127 12:43:09.568789 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:43:09 crc kubenswrapper[4796]: E1127 12:43:09.569624 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:43:14 crc kubenswrapper[4796]: I1127 12:43:14.044516 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8wmkg_15b7bb22-0ae1-46af-a1cf-4024996c996f/control-plane-machine-set-operator/0.log" Nov 27 12:43:14 crc kubenswrapper[4796]: I1127 12:43:14.534802 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bxf5l_96d741c5-ebc0-4c04-9192-c2661a5d0e0a/machine-api-operator/0.log" Nov 27 12:43:14 crc kubenswrapper[4796]: I1127 12:43:14.542095 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bxf5l_96d741c5-ebc0-4c04-9192-c2661a5d0e0a/kube-rbac-proxy/0.log" Nov 27 12:43:23 crc kubenswrapper[4796]: I1127 12:43:23.578414 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:43:23 crc kubenswrapper[4796]: E1127 12:43:23.579253 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:43:25 crc kubenswrapper[4796]: I1127 12:43:25.109799 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nx48n_22e38aa1-4d12-4b1f-bfef-12d303bc12ca/cert-manager-controller/0.log" Nov 27 12:43:25 crc kubenswrapper[4796]: I1127 12:43:25.280604 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-scllt_2a72ae09-3cbb-4ab3-91b3-0aab83b57b74/cert-manager-cainjector/0.log" Nov 27 12:43:25 crc kubenswrapper[4796]: I1127 12:43:25.323294 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qrd89_bfed3074-7dee-4f77-900a-8546651ce2e6/cert-manager-webhook/0.log" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.414365 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g8jfz_68dd98a6-a2d9-4a74-81c4-fab3ae549244/nmstate-console-plugin/0.log" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.568766 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:43:37 crc kubenswrapper[4796]: E1127 12:43:37.569118 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.580502 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vtgx5_e3cc1a65-8f8a-477c-96f1-3cd385bfa910/nmstate-handler/0.log" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.604462 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2tbf2_72f7a9b6-d5a1-4b35-8221-8c7afcaa203c/kube-rbac-proxy/0.log" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.613137 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2tbf2_72f7a9b6-d5a1-4b35-8221-8c7afcaa203c/nmstate-metrics/0.log" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.778437 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-l6tgw_2e5faff6-d103-4aa7-ae65-77abe90548c5/nmstate-operator/0.log" Nov 27 12:43:37 crc kubenswrapper[4796]: I1127 12:43:37.845815 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-g7rhk_44d16d5c-7e22-4bf4-824e-9d58d750ac7d/nmstate-webhook/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.066742 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-f8gw9_f8304148-f799-4ea3-880c-7d7e5b744cef/kube-rbac-proxy/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.141575 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-f8gw9_f8304148-f799-4ea3-880c-7d7e5b744cef/controller/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.273460 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.447284 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.468825 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.499354 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.519882 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.637904 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.650800 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.672497 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.697660 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.868229 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.884990 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.888688 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:43:51 crc kubenswrapper[4796]: I1127 12:43:51.915717 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/controller/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.056462 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/frr-metrics/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.095842 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/kube-rbac-proxy/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.155813 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/kube-rbac-proxy-frr/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.286537 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/reloader/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.385705 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-85888_10c4179c-1abe-40a0-a5eb-ea7f70a4fde8/frr-k8s-webhook-server/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.488227 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-68bdcff486-mtshf_e27fc43a-060d-4f5f-9f52-4f75c9e4b651/manager/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.569059 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:43:52 crc kubenswrapper[4796]: E1127 12:43:52.569522 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.678381 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-57ffb79475-qkvck_c1a62d95-1792-4779-8f23-da2e94806d72/webhook-server/0.log" Nov 27 12:43:52 crc kubenswrapper[4796]: I1127 12:43:52.860776 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nd74_d6863326-623f-4817-ba2b-65f8b1887f5d/kube-rbac-proxy/0.log" Nov 27 12:43:53 crc kubenswrapper[4796]: I1127 12:43:53.493443 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nd74_d6863326-623f-4817-ba2b-65f8b1887f5d/speaker/0.log" Nov 27 12:43:53 crc kubenswrapper[4796]: I1127 12:43:53.714144 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/frr/0.log" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.347320 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dk7db"] Nov 27 12:43:58 crc kubenswrapper[4796]: E1127 12:43:58.348250 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2fc84c-1468-4627-b72b-47f9b2f5d630" containerName="container-00" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.348276 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2fc84c-1468-4627-b72b-47f9b2f5d630" containerName="container-00" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.348531 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2fc84c-1468-4627-b72b-47f9b2f5d630" containerName="container-00" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.350410 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.354507 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dk7db"] Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.511120 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-catalog-content\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.511398 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-utilities\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.511615 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-kube-api-access-5hpr8\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.612982 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-kube-api-access-5hpr8\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.613646 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-catalog-content\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.613773 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-utilities\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.614207 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-catalog-content\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.614283 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-utilities\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.640240 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-kube-api-access-5hpr8\") pod \"community-operators-dk7db\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:58 crc kubenswrapper[4796]: I1127 12:43:58.679892 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:43:59 crc kubenswrapper[4796]: I1127 12:43:59.168891 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dk7db"] Nov 27 12:43:59 crc kubenswrapper[4796]: I1127 12:43:59.193900 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerStarted","Data":"e1fe621717dec0ede0000cec1f9e4a00e73e7d0245a100c04ee6446def3cebcf"} Nov 27 12:44:00 crc kubenswrapper[4796]: I1127 12:44:00.207367 4796 generic.go:334] "Generic (PLEG): container finished" podID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerID="7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373" exitCode=0 Nov 27 12:44:00 crc kubenswrapper[4796]: I1127 12:44:00.207470 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerDied","Data":"7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373"} Nov 27 12:44:00 crc kubenswrapper[4796]: I1127 12:44:00.211914 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:44:02 crc kubenswrapper[4796]: I1127 12:44:02.243611 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerStarted","Data":"ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99"} Nov 27 12:44:03 crc kubenswrapper[4796]: I1127 12:44:03.258425 4796 generic.go:334] "Generic (PLEG): container finished" podID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerID="ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99" exitCode=0 Nov 27 12:44:03 crc kubenswrapper[4796]: I1127 12:44:03.258474 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerDied","Data":"ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99"} Nov 27 12:44:05 crc kubenswrapper[4796]: I1127 12:44:05.279886 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerStarted","Data":"feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b"} Nov 27 12:44:05 crc kubenswrapper[4796]: I1127 12:44:05.305865 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dk7db" podStartSLOduration=3.4166098480000002 podStartE2EDuration="7.305843215s" podCreationTimestamp="2025-11-27 12:43:58 +0000 UTC" firstStartedPulling="2025-11-27 12:44:00.211398113 +0000 UTC m=+4757.729717061" lastFinishedPulling="2025-11-27 12:44:04.10063148 +0000 UTC m=+4761.618950428" observedRunningTime="2025-11-27 12:44:05.302708288 +0000 UTC m=+4762.821027216" watchObservedRunningTime="2025-11-27 12:44:05.305843215 +0000 UTC m=+4762.824162153" Nov 27 12:44:06 crc kubenswrapper[4796]: I1127 12:44:06.995771 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/util/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.188467 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/pull/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.188614 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/pull/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.366429 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/extract/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.403522 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/pull/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.424334 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/util/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.433743 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/util/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.565033 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/util/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.569039 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:44:07 crc kubenswrapper[4796]: E1127 12:44:07.569301 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.733557 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/util/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.734702 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/pull/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.735731 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/pull/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.894879 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/extract/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.894968 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/pull/0.log" Nov 27 12:44:07 crc kubenswrapper[4796]: I1127 12:44:07.917517 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/util/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.058423 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/util/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.241698 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/pull/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.260540 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/pull/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.265707 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/util/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.471121 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/extract/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.478017 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/util/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.505668 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/pull/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.634409 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-utilities/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.680159 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.680216 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.790708 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-utilities/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.816931 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-content/0.log" Nov 27 12:44:08 crc kubenswrapper[4796]: I1127 12:44:08.823880 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-content/0.log" Nov 27 12:44:09 crc kubenswrapper[4796]: I1127 12:44:09.425308 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:44:09 crc kubenswrapper[4796]: I1127 12:44:09.472323 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:44:09 crc kubenswrapper[4796]: I1127 12:44:09.576483 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-content/0.log" Nov 27 12:44:09 crc kubenswrapper[4796]: I1127 12:44:09.632321 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-utilities/0.log" Nov 27 12:44:09 crc kubenswrapper[4796]: I1127 12:44:09.669441 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dk7db"] Nov 27 12:44:09 crc kubenswrapper[4796]: I1127 12:44:09.787246 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/extract-utilities/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.091321 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/extract-utilities/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.093685 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/extract-content/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.139887 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/extract-content/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.145122 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/registry-server/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.274541 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/extract-content/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.276332 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/extract-utilities/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.330915 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dk7db_c4bbe898-d010-49c3-9a96-dedcb1f5d65b/registry-server/0.log" Nov 27 12:44:10 crc kubenswrapper[4796]: I1127 12:44:10.484508 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-utilities/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.257705 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-content/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.323560 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-content/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.326833 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-utilities/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.333108 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dk7db" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="registry-server" containerID="cri-o://feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b" gracePeriod=2 Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.615522 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-utilities/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.615553 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-content/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.690169 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-pnvmc_4a5aa833-5ce1-4847-bd66-45dcae6a2015/marketplace-operator/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.874928 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-utilities/0.log" Nov 27 12:44:11 crc kubenswrapper[4796]: I1127 12:44:11.997020 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.024317 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-content/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.032687 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-utilities/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.074485 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-content/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.089698 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-utilities\") pod \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.089798 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-kube-api-access-5hpr8\") pod \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.089833 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-catalog-content\") pod \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\" (UID: \"c4bbe898-d010-49c3-9a96-dedcb1f5d65b\") " Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.090607 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-utilities" (OuterVolumeSpecName: "utilities") pod "c4bbe898-d010-49c3-9a96-dedcb1f5d65b" (UID: "c4bbe898-d010-49c3-9a96-dedcb1f5d65b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.095426 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-kube-api-access-5hpr8" (OuterVolumeSpecName: "kube-api-access-5hpr8") pod "c4bbe898-d010-49c3-9a96-dedcb1f5d65b" (UID: "c4bbe898-d010-49c3-9a96-dedcb1f5d65b"). InnerVolumeSpecName "kube-api-access-5hpr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.138757 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4bbe898-d010-49c3-9a96-dedcb1f5d65b" (UID: "c4bbe898-d010-49c3-9a96-dedcb1f5d65b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.191988 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.192206 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hpr8\" (UniqueName: \"kubernetes.io/projected/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-kube-api-access-5hpr8\") on node \"crc\" DevicePath \"\"" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.192217 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4bbe898-d010-49c3-9a96-dedcb1f5d65b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.233608 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/registry-server/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.267168 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-content/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.300560 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-utilities/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.343816 4796 generic.go:334] "Generic (PLEG): container finished" podID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerID="feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b" exitCode=0 Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.343860 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerDied","Data":"feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b"} Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.343879 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dk7db" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.343891 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dk7db" event={"ID":"c4bbe898-d010-49c3-9a96-dedcb1f5d65b","Type":"ContainerDied","Data":"e1fe621717dec0ede0000cec1f9e4a00e73e7d0245a100c04ee6446def3cebcf"} Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.343927 4796 scope.go:117] "RemoveContainer" containerID="feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.363626 4796 scope.go:117] "RemoveContainer" containerID="ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.395210 4796 scope.go:117] "RemoveContainer" containerID="7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.398261 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dk7db"] Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.409949 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dk7db"] Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.428042 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/extract-utilities/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.444144 4796 scope.go:117] "RemoveContainer" containerID="feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b" Nov 27 12:44:12 crc kubenswrapper[4796]: E1127 12:44:12.444533 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b\": container with ID starting with feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b not found: ID does not exist" containerID="feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.444569 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b"} err="failed to get container status \"feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b\": rpc error: code = NotFound desc = could not find container \"feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b\": container with ID starting with feda1373343f48a2d1c22874543fb396e5aedb319d8b44fe3e271b3e8814545b not found: ID does not exist" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.444589 4796 scope.go:117] "RemoveContainer" containerID="ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99" Nov 27 12:44:12 crc kubenswrapper[4796]: E1127 12:44:12.444827 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99\": container with ID starting with ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99 not found: ID does not exist" containerID="ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.444852 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99"} err="failed to get container status \"ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99\": rpc error: code = NotFound desc = could not find container \"ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99\": container with ID starting with ef81f226bb16645fd6bda2f0bf1a2e9306c49aca6a8effb87faae9b7e392cd99 not found: ID does not exist" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.444867 4796 scope.go:117] "RemoveContainer" containerID="7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373" Nov 27 12:44:12 crc kubenswrapper[4796]: E1127 12:44:12.445088 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373\": container with ID starting with 7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373 not found: ID does not exist" containerID="7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.445110 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373"} err="failed to get container status \"7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373\": rpc error: code = NotFound desc = could not find container \"7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373\": container with ID starting with 7ce1232b3a3c1c2eb4a31465776930e4ab80d3ec471d22d56635f1698aaa0373 not found: ID does not exist" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.537407 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/registry-server/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.584050 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/extract-utilities/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.619377 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/extract-content/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.623135 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/extract-content/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.766637 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/extract-content/0.log" Nov 27 12:44:12 crc kubenswrapper[4796]: I1127 12:44:12.768165 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/extract-utilities/0.log" Nov 27 12:44:13 crc kubenswrapper[4796]: I1127 12:44:13.323700 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-b5fl4_e9abb298-96b8-41ea-b357-534bc4947657/registry-server/0.log" Nov 27 12:44:13 crc kubenswrapper[4796]: I1127 12:44:13.588863 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" path="/var/lib/kubelet/pods/c4bbe898-d010-49c3-9a96-dedcb1f5d65b/volumes" Nov 27 12:44:18 crc kubenswrapper[4796]: I1127 12:44:18.570071 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:44:18 crc kubenswrapper[4796]: E1127 12:44:18.570866 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:44:24 crc kubenswrapper[4796]: I1127 12:44:24.886495 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-q7hl9_12aebb1f-5956-4364-b2c9-80cf77ddb77a/prometheus-operator/0.log" Nov 27 12:44:25 crc kubenswrapper[4796]: I1127 12:44:25.082481 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67d796579b-bbzgw_e2638406-4d7c-404f-9906-544ce502974b/prometheus-operator-admission-webhook/0.log" Nov 27 12:44:25 crc kubenswrapper[4796]: I1127 12:44:25.269691 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67d796579b-n5622_bd9f4f30-7f8c-40d0-b559-368b90f90b4b/prometheus-operator-admission-webhook/0.log" Nov 27 12:44:25 crc kubenswrapper[4796]: I1127 12:44:25.355277 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-x26pv_2b928049-ffbb-4f1a-9d05-14f4bae23427/operator/0.log" Nov 27 12:44:25 crc kubenswrapper[4796]: I1127 12:44:25.474777 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-t8f4v_94471100-b73d-4ca3-be61-4eac947eae64/perses-operator/0.log" Nov 27 12:44:31 crc kubenswrapper[4796]: I1127 12:44:31.569389 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:44:31 crc kubenswrapper[4796]: E1127 12:44:31.570360 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:44:44 crc kubenswrapper[4796]: I1127 12:44:44.568864 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:44:44 crc kubenswrapper[4796]: E1127 12:44:44.569724 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:44:59 crc kubenswrapper[4796]: I1127 12:44:59.570050 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:44:59 crc kubenswrapper[4796]: E1127 12:44:59.571104 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.174135 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz"] Nov 27 12:45:00 crc kubenswrapper[4796]: E1127 12:45:00.174547 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="registry-server" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.174572 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="registry-server" Nov 27 12:45:00 crc kubenswrapper[4796]: E1127 12:45:00.174584 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="extract-content" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.174590 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="extract-content" Nov 27 12:45:00 crc kubenswrapper[4796]: E1127 12:45:00.174632 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="extract-utilities" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.174638 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="extract-utilities" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.174823 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bbe898-d010-49c3-9a96-dedcb1f5d65b" containerName="registry-server" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.175572 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.179833 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.180094 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.185564 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz"] Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.322870 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/697ecd16-530b-48cf-97c6-89539686da05-config-volume\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.322957 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5x9n\" (UniqueName: \"kubernetes.io/projected/697ecd16-530b-48cf-97c6-89539686da05-kube-api-access-t5x9n\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.323084 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/697ecd16-530b-48cf-97c6-89539686da05-secret-volume\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.424943 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/697ecd16-530b-48cf-97c6-89539686da05-config-volume\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.425063 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5x9n\" (UniqueName: \"kubernetes.io/projected/697ecd16-530b-48cf-97c6-89539686da05-kube-api-access-t5x9n\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.425334 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/697ecd16-530b-48cf-97c6-89539686da05-secret-volume\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.426063 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/697ecd16-530b-48cf-97c6-89539686da05-config-volume\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.444166 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/697ecd16-530b-48cf-97c6-89539686da05-secret-volume\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.451109 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5x9n\" (UniqueName: \"kubernetes.io/projected/697ecd16-530b-48cf-97c6-89539686da05-kube-api-access-t5x9n\") pod \"collect-profiles-29404125-rrsjz\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.501572 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:00 crc kubenswrapper[4796]: I1127 12:45:00.970396 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz"] Nov 27 12:45:01 crc kubenswrapper[4796]: I1127 12:45:01.841848 4796 generic.go:334] "Generic (PLEG): container finished" podID="697ecd16-530b-48cf-97c6-89539686da05" containerID="5b23b808404b39987bdc7085e0db22b24433372d3ad0eadf8b5c9638ccacbe4d" exitCode=0 Nov 27 12:45:01 crc kubenswrapper[4796]: I1127 12:45:01.841913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" event={"ID":"697ecd16-530b-48cf-97c6-89539686da05","Type":"ContainerDied","Data":"5b23b808404b39987bdc7085e0db22b24433372d3ad0eadf8b5c9638ccacbe4d"} Nov 27 12:45:01 crc kubenswrapper[4796]: I1127 12:45:01.842678 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" event={"ID":"697ecd16-530b-48cf-97c6-89539686da05","Type":"ContainerStarted","Data":"1f537b98384cef3c349fd877a3c2e887ab6be78326b959b874494e3b0b56da60"} Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.264671 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.394397 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5x9n\" (UniqueName: \"kubernetes.io/projected/697ecd16-530b-48cf-97c6-89539686da05-kube-api-access-t5x9n\") pod \"697ecd16-530b-48cf-97c6-89539686da05\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.394672 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/697ecd16-530b-48cf-97c6-89539686da05-secret-volume\") pod \"697ecd16-530b-48cf-97c6-89539686da05\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.394920 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/697ecd16-530b-48cf-97c6-89539686da05-config-volume\") pod \"697ecd16-530b-48cf-97c6-89539686da05\" (UID: \"697ecd16-530b-48cf-97c6-89539686da05\") " Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.396024 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/697ecd16-530b-48cf-97c6-89539686da05-config-volume" (OuterVolumeSpecName: "config-volume") pod "697ecd16-530b-48cf-97c6-89539686da05" (UID: "697ecd16-530b-48cf-97c6-89539686da05"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.400505 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/697ecd16-530b-48cf-97c6-89539686da05-kube-api-access-t5x9n" (OuterVolumeSpecName: "kube-api-access-t5x9n") pod "697ecd16-530b-48cf-97c6-89539686da05" (UID: "697ecd16-530b-48cf-97c6-89539686da05"). InnerVolumeSpecName "kube-api-access-t5x9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.402014 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697ecd16-530b-48cf-97c6-89539686da05-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "697ecd16-530b-48cf-97c6-89539686da05" (UID: "697ecd16-530b-48cf-97c6-89539686da05"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.497610 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5x9n\" (UniqueName: \"kubernetes.io/projected/697ecd16-530b-48cf-97c6-89539686da05-kube-api-access-t5x9n\") on node \"crc\" DevicePath \"\"" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.497663 4796 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/697ecd16-530b-48cf-97c6-89539686da05-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.497673 4796 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/697ecd16-530b-48cf-97c6-89539686da05-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.870339 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" event={"ID":"697ecd16-530b-48cf-97c6-89539686da05","Type":"ContainerDied","Data":"1f537b98384cef3c349fd877a3c2e887ab6be78326b959b874494e3b0b56da60"} Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.870414 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f537b98384cef3c349fd877a3c2e887ab6be78326b959b874494e3b0b56da60" Nov 27 12:45:03 crc kubenswrapper[4796]: I1127 12:45:03.870434 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404125-rrsjz" Nov 27 12:45:04 crc kubenswrapper[4796]: I1127 12:45:04.363060 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5"] Nov 27 12:45:04 crc kubenswrapper[4796]: I1127 12:45:04.379105 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404080-vh5j5"] Nov 27 12:45:05 crc kubenswrapper[4796]: I1127 12:45:05.584561 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4228ea4e-2713-4d86-8922-8e4e5e4fbdbe" path="/var/lib/kubelet/pods/4228ea4e-2713-4d86-8922-8e4e5e4fbdbe/volumes" Nov 27 12:45:11 crc kubenswrapper[4796]: I1127 12:45:11.571346 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:45:11 crc kubenswrapper[4796]: E1127 12:45:11.572141 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:45:24 crc kubenswrapper[4796]: I1127 12:45:24.569517 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:45:24 crc kubenswrapper[4796]: E1127 12:45:24.570641 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:45:37 crc kubenswrapper[4796]: I1127 12:45:37.569773 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:45:37 crc kubenswrapper[4796]: E1127 12:45:37.570714 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:45:51 crc kubenswrapper[4796]: I1127 12:45:51.569464 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:45:51 crc kubenswrapper[4796]: E1127 12:45:51.570335 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:45:54 crc kubenswrapper[4796]: I1127 12:45:54.770389 4796 scope.go:117] "RemoveContainer" containerID="c819fb7b4d9217b8e4b5242ed7e2363d0dca8ff8737017affafebe3172cba86e" Nov 27 12:45:56 crc kubenswrapper[4796]: I1127 12:45:56.472945 4796 generic.go:334] "Generic (PLEG): container finished" podID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerID="0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70" exitCode=0 Nov 27 12:45:56 crc kubenswrapper[4796]: I1127 12:45:56.473348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-76h9f/must-gather-62n49" event={"ID":"907cb4cc-62e8-45fa-88a1-5226e75b11d2","Type":"ContainerDied","Data":"0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70"} Nov 27 12:45:56 crc kubenswrapper[4796]: I1127 12:45:56.474334 4796 scope.go:117] "RemoveContainer" containerID="0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70" Nov 27 12:45:56 crc kubenswrapper[4796]: I1127 12:45:56.642756 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-76h9f_must-gather-62n49_907cb4cc-62e8-45fa-88a1-5226e75b11d2/gather/0.log" Nov 27 12:46:02 crc kubenswrapper[4796]: I1127 12:46:02.569253 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:46:02 crc kubenswrapper[4796]: E1127 12:46:02.570113 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:46:04 crc kubenswrapper[4796]: I1127 12:46:04.611073 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-76h9f/must-gather-62n49"] Nov 27 12:46:04 crc kubenswrapper[4796]: I1127 12:46:04.611730 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-76h9f/must-gather-62n49" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="copy" containerID="cri-o://1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4" gracePeriod=2 Nov 27 12:46:04 crc kubenswrapper[4796]: I1127 12:46:04.620887 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-76h9f/must-gather-62n49"] Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.059255 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-76h9f_must-gather-62n49_907cb4cc-62e8-45fa-88a1-5226e75b11d2/copy/0.log" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.059897 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.236824 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/907cb4cc-62e8-45fa-88a1-5226e75b11d2-must-gather-output\") pod \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.236921 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2t47\" (UniqueName: \"kubernetes.io/projected/907cb4cc-62e8-45fa-88a1-5226e75b11d2-kube-api-access-q2t47\") pod \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\" (UID: \"907cb4cc-62e8-45fa-88a1-5226e75b11d2\") " Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.242625 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907cb4cc-62e8-45fa-88a1-5226e75b11d2-kube-api-access-q2t47" (OuterVolumeSpecName: "kube-api-access-q2t47") pod "907cb4cc-62e8-45fa-88a1-5226e75b11d2" (UID: "907cb4cc-62e8-45fa-88a1-5226e75b11d2"). InnerVolumeSpecName "kube-api-access-q2t47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.339498 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2t47\" (UniqueName: \"kubernetes.io/projected/907cb4cc-62e8-45fa-88a1-5226e75b11d2-kube-api-access-q2t47\") on node \"crc\" DevicePath \"\"" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.378395 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/907cb4cc-62e8-45fa-88a1-5226e75b11d2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "907cb4cc-62e8-45fa-88a1-5226e75b11d2" (UID: "907cb4cc-62e8-45fa-88a1-5226e75b11d2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.441436 4796 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/907cb4cc-62e8-45fa-88a1-5226e75b11d2-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.572004 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-76h9f_must-gather-62n49_907cb4cc-62e8-45fa-88a1-5226e75b11d2/copy/0.log" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.572500 4796 generic.go:334] "Generic (PLEG): container finished" podID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerID="1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4" exitCode=143 Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.572641 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-76h9f/must-gather-62n49" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.593915 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" path="/var/lib/kubelet/pods/907cb4cc-62e8-45fa-88a1-5226e75b11d2/volumes" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.597071 4796 scope.go:117] "RemoveContainer" containerID="1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.623833 4796 scope.go:117] "RemoveContainer" containerID="0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.682087 4796 scope.go:117] "RemoveContainer" containerID="1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4" Nov 27 12:46:05 crc kubenswrapper[4796]: E1127 12:46:05.691892 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4\": container with ID starting with 1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4 not found: ID does not exist" containerID="1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.691957 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4"} err="failed to get container status \"1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4\": rpc error: code = NotFound desc = could not find container \"1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4\": container with ID starting with 1ea290d4359afb6649f78630e110f524c2e344ffd7c3e20f7c094e5e677912c4 not found: ID does not exist" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.691996 4796 scope.go:117] "RemoveContainer" containerID="0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70" Nov 27 12:46:05 crc kubenswrapper[4796]: E1127 12:46:05.692682 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70\": container with ID starting with 0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70 not found: ID does not exist" containerID="0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70" Nov 27 12:46:05 crc kubenswrapper[4796]: I1127 12:46:05.692721 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70"} err="failed to get container status \"0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70\": rpc error: code = NotFound desc = could not find container \"0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70\": container with ID starting with 0f1dbf2f6e83804d254a8d33197b755f223edc0df22cb0df4ebe8f600f017b70 not found: ID does not exist" Nov 27 12:46:16 crc kubenswrapper[4796]: I1127 12:46:16.569291 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:46:16 crc kubenswrapper[4796]: E1127 12:46:16.570074 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:46:28 crc kubenswrapper[4796]: I1127 12:46:28.569953 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:46:28 crc kubenswrapper[4796]: E1127 12:46:28.570852 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:46:39 crc kubenswrapper[4796]: I1127 12:46:39.569651 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:46:39 crc kubenswrapper[4796]: I1127 12:46:39.921768 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"cf308b662ff0de97573e44404d006635d7780706d2dd1c7c90a2c1aba1c97d5c"} Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.439716 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2m27c"] Nov 27 12:47:43 crc kubenswrapper[4796]: E1127 12:47:43.440747 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="gather" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.440763 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="gather" Nov 27 12:47:43 crc kubenswrapper[4796]: E1127 12:47:43.440801 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="copy" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.440809 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="copy" Nov 27 12:47:43 crc kubenswrapper[4796]: E1127 12:47:43.440828 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="697ecd16-530b-48cf-97c6-89539686da05" containerName="collect-profiles" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.440835 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="697ecd16-530b-48cf-97c6-89539686da05" containerName="collect-profiles" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.441053 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="gather" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.441065 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="697ecd16-530b-48cf-97c6-89539686da05" containerName="collect-profiles" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.441099 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="907cb4cc-62e8-45fa-88a1-5226e75b11d2" containerName="copy" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.443006 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.451751 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2m27c"] Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.639328 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e381c75-af72-45c2-a671-b8c506d16fb6-utilities\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.639471 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e381c75-af72-45c2-a671-b8c506d16fb6-catalog-content\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.639617 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmgmk\" (UniqueName: \"kubernetes.io/projected/9e381c75-af72-45c2-a671-b8c506d16fb6-kube-api-access-cmgmk\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.741113 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmgmk\" (UniqueName: \"kubernetes.io/projected/9e381c75-af72-45c2-a671-b8c506d16fb6-kube-api-access-cmgmk\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.741243 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e381c75-af72-45c2-a671-b8c506d16fb6-utilities\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.741339 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e381c75-af72-45c2-a671-b8c506d16fb6-catalog-content\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.741815 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e381c75-af72-45c2-a671-b8c506d16fb6-utilities\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.741823 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e381c75-af72-45c2-a671-b8c506d16fb6-catalog-content\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.763533 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmgmk\" (UniqueName: \"kubernetes.io/projected/9e381c75-af72-45c2-a671-b8c506d16fb6-kube-api-access-cmgmk\") pod \"redhat-operators-2m27c\" (UID: \"9e381c75-af72-45c2-a671-b8c506d16fb6\") " pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:43 crc kubenswrapper[4796]: I1127 12:47:43.772513 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:44 crc kubenswrapper[4796]: W1127 12:47:44.243846 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e381c75_af72_45c2_a671_b8c506d16fb6.slice/crio-150f3fcb60d0d28d1ffad8522b857e30f6489c523b5e294ad39f8a8873b1797c WatchSource:0}: Error finding container 150f3fcb60d0d28d1ffad8522b857e30f6489c523b5e294ad39f8a8873b1797c: Status 404 returned error can't find the container with id 150f3fcb60d0d28d1ffad8522b857e30f6489c523b5e294ad39f8a8873b1797c Nov 27 12:47:44 crc kubenswrapper[4796]: I1127 12:47:44.250330 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2m27c"] Nov 27 12:47:44 crc kubenswrapper[4796]: I1127 12:47:44.613435 4796 generic.go:334] "Generic (PLEG): container finished" podID="9e381c75-af72-45c2-a671-b8c506d16fb6" containerID="1f4adc94acaa848396a01a5afb275b4cb750a45ba43f6868748e44890d4240fe" exitCode=0 Nov 27 12:47:44 crc kubenswrapper[4796]: I1127 12:47:44.613553 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27c" event={"ID":"9e381c75-af72-45c2-a671-b8c506d16fb6","Type":"ContainerDied","Data":"1f4adc94acaa848396a01a5afb275b4cb750a45ba43f6868748e44890d4240fe"} Nov 27 12:47:44 crc kubenswrapper[4796]: I1127 12:47:44.613727 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27c" event={"ID":"9e381c75-af72-45c2-a671-b8c506d16fb6","Type":"ContainerStarted","Data":"150f3fcb60d0d28d1ffad8522b857e30f6489c523b5e294ad39f8a8873b1797c"} Nov 27 12:47:51 crc kubenswrapper[4796]: I1127 12:47:51.696476 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27c" event={"ID":"9e381c75-af72-45c2-a671-b8c506d16fb6","Type":"ContainerStarted","Data":"b9bb4b42207c3a9b42bc799417cd0f60dcd34f517423b8129c121898e0a0a0b1"} Nov 27 12:47:52 crc kubenswrapper[4796]: I1127 12:47:52.712080 4796 generic.go:334] "Generic (PLEG): container finished" podID="9e381c75-af72-45c2-a671-b8c506d16fb6" containerID="b9bb4b42207c3a9b42bc799417cd0f60dcd34f517423b8129c121898e0a0a0b1" exitCode=0 Nov 27 12:47:52 crc kubenswrapper[4796]: I1127 12:47:52.712163 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27c" event={"ID":"9e381c75-af72-45c2-a671-b8c506d16fb6","Type":"ContainerDied","Data":"b9bb4b42207c3a9b42bc799417cd0f60dcd34f517423b8129c121898e0a0a0b1"} Nov 27 12:47:53 crc kubenswrapper[4796]: I1127 12:47:53.734286 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27c" event={"ID":"9e381c75-af72-45c2-a671-b8c506d16fb6","Type":"ContainerStarted","Data":"b6702871b4f7c55421fed4887501ad39aa1b2a96a2c7d35a8409b937e7744938"} Nov 27 12:47:53 crc kubenswrapper[4796]: I1127 12:47:53.768699 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2m27c" podStartSLOduration=1.904395998 podStartE2EDuration="10.768680095s" podCreationTimestamp="2025-11-27 12:47:43 +0000 UTC" firstStartedPulling="2025-11-27 12:47:44.61526179 +0000 UTC m=+4982.133580708" lastFinishedPulling="2025-11-27 12:47:53.479545877 +0000 UTC m=+4990.997864805" observedRunningTime="2025-11-27 12:47:53.754461517 +0000 UTC m=+4991.272780455" watchObservedRunningTime="2025-11-27 12:47:53.768680095 +0000 UTC m=+4991.286999013" Nov 27 12:47:53 crc kubenswrapper[4796]: I1127 12:47:53.773569 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:53 crc kubenswrapper[4796]: I1127 12:47:53.773778 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:47:54 crc kubenswrapper[4796]: I1127 12:47:54.817844 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2m27c" podUID="9e381c75-af72-45c2-a671-b8c506d16fb6" containerName="registry-server" probeResult="failure" output=< Nov 27 12:47:54 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:47:54 crc kubenswrapper[4796]: > Nov 27 12:47:54 crc kubenswrapper[4796]: I1127 12:47:54.882065 4796 scope.go:117] "RemoveContainer" containerID="bbf2fe8905fcfa81f86809192981816b48a96bbe29771b1d25488c1ea2050c4e" Nov 27 12:48:03 crc kubenswrapper[4796]: I1127 12:48:03.854850 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:48:03 crc kubenswrapper[4796]: I1127 12:48:03.934391 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2m27c" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.011291 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2m27c"] Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.098483 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5fl4"] Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.099013 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b5fl4" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="registry-server" containerID="cri-o://ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023" gracePeriod=2 Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.569198 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.662664 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-utilities\") pod \"e9abb298-96b8-41ea-b357-534bc4947657\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.662742 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-catalog-content\") pod \"e9abb298-96b8-41ea-b357-534bc4947657\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.662795 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx9mt\" (UniqueName: \"kubernetes.io/projected/e9abb298-96b8-41ea-b357-534bc4947657-kube-api-access-xx9mt\") pod \"e9abb298-96b8-41ea-b357-534bc4947657\" (UID: \"e9abb298-96b8-41ea-b357-534bc4947657\") " Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.664503 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-utilities" (OuterVolumeSpecName: "utilities") pod "e9abb298-96b8-41ea-b357-534bc4947657" (UID: "e9abb298-96b8-41ea-b357-534bc4947657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.672522 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9abb298-96b8-41ea-b357-534bc4947657-kube-api-access-xx9mt" (OuterVolumeSpecName: "kube-api-access-xx9mt") pod "e9abb298-96b8-41ea-b357-534bc4947657" (UID: "e9abb298-96b8-41ea-b357-534bc4947657"). InnerVolumeSpecName "kube-api-access-xx9mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.765843 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.765887 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx9mt\" (UniqueName: \"kubernetes.io/projected/e9abb298-96b8-41ea-b357-534bc4947657-kube-api-access-xx9mt\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.777146 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9abb298-96b8-41ea-b357-534bc4947657" (UID: "e9abb298-96b8-41ea-b357-534bc4947657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.849076 4796 generic.go:334] "Generic (PLEG): container finished" podID="e9abb298-96b8-41ea-b357-534bc4947657" containerID="ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023" exitCode=0 Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.849166 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerDied","Data":"ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023"} Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.849224 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5fl4" event={"ID":"e9abb298-96b8-41ea-b357-534bc4947657","Type":"ContainerDied","Data":"43133b8766e6acd7c61c39155f81c72a67e2272acabc9b6de6e319c520178bb1"} Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.849290 4796 scope.go:117] "RemoveContainer" containerID="ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.849538 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5fl4" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.866709 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9abb298-96b8-41ea-b357-534bc4947657-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.886342 4796 scope.go:117] "RemoveContainer" containerID="dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab" Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.889847 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5fl4"] Nov 27 12:48:04 crc kubenswrapper[4796]: I1127 12:48:04.902348 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b5fl4"] Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.386710 4796 scope.go:117] "RemoveContainer" containerID="0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.440693 4796 scope.go:117] "RemoveContainer" containerID="ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023" Nov 27 12:48:05 crc kubenswrapper[4796]: E1127 12:48:05.441290 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023\": container with ID starting with ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023 not found: ID does not exist" containerID="ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.441341 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023"} err="failed to get container status \"ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023\": rpc error: code = NotFound desc = could not find container \"ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023\": container with ID starting with ea7a386b6234af7a97f5922848d20be0b4c83327066acda7f401d01d51f8a023 not found: ID does not exist" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.441369 4796 scope.go:117] "RemoveContainer" containerID="dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab" Nov 27 12:48:05 crc kubenswrapper[4796]: E1127 12:48:05.441753 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab\": container with ID starting with dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab not found: ID does not exist" containerID="dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.441802 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab"} err="failed to get container status \"dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab\": rpc error: code = NotFound desc = could not find container \"dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab\": container with ID starting with dcfafa94fbc3e5ae93e83463813635b4f62539f269c935e4c7f283b5097ddeab not found: ID does not exist" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.441834 4796 scope.go:117] "RemoveContainer" containerID="0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17" Nov 27 12:48:05 crc kubenswrapper[4796]: E1127 12:48:05.442186 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17\": container with ID starting with 0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17 not found: ID does not exist" containerID="0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.442220 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17"} err="failed to get container status \"0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17\": rpc error: code = NotFound desc = could not find container \"0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17\": container with ID starting with 0a4c4c611dba1ff80167e6c558b6b1c405151030c8d77e041e08f9b5f16c1f17 not found: ID does not exist" Nov 27 12:48:05 crc kubenswrapper[4796]: I1127 12:48:05.602905 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9abb298-96b8-41ea-b357-534bc4947657" path="/var/lib/kubelet/pods/e9abb298-96b8-41ea-b357-534bc4947657/volumes" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.030708 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xtg9m"] Nov 27 12:48:27 crc kubenswrapper[4796]: E1127 12:48:27.032652 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="extract-utilities" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.032969 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="extract-utilities" Nov 27 12:48:27 crc kubenswrapper[4796]: E1127 12:48:27.033172 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="extract-content" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.033206 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="extract-content" Nov 27 12:48:27 crc kubenswrapper[4796]: E1127 12:48:27.033302 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="registry-server" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.033322 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="registry-server" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.037156 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9abb298-96b8-41ea-b357-534bc4947657" containerName="registry-server" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.043395 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.046367 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvqzs\" (UniqueName: \"kubernetes.io/projected/eef7b83c-fe8c-4efa-8658-34555eab03c8-kube-api-access-bvqzs\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.046420 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-utilities\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.046499 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-catalog-content\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.048924 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xtg9m"] Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.147934 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvqzs\" (UniqueName: \"kubernetes.io/projected/eef7b83c-fe8c-4efa-8658-34555eab03c8-kube-api-access-bvqzs\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.147997 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-utilities\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.148079 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-catalog-content\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.148808 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-utilities\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.148832 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-catalog-content\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.174227 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvqzs\" (UniqueName: \"kubernetes.io/projected/eef7b83c-fe8c-4efa-8658-34555eab03c8-kube-api-access-bvqzs\") pod \"certified-operators-xtg9m\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.368990 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:27 crc kubenswrapper[4796]: I1127 12:48:27.894621 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xtg9m"] Nov 27 12:48:28 crc kubenswrapper[4796]: I1127 12:48:28.116831 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerStarted","Data":"9cee71173bb458d8f5eb887863097ffe53e74f400499b2df08d7c30978482ad0"} Nov 27 12:48:29 crc kubenswrapper[4796]: I1127 12:48:29.128812 4796 generic.go:334] "Generic (PLEG): container finished" podID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerID="9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7" exitCode=0 Nov 27 12:48:29 crc kubenswrapper[4796]: I1127 12:48:29.128919 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerDied","Data":"9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7"} Nov 27 12:48:30 crc kubenswrapper[4796]: I1127 12:48:30.140372 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerStarted","Data":"a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b"} Nov 27 12:48:32 crc kubenswrapper[4796]: I1127 12:48:32.160971 4796 generic.go:334] "Generic (PLEG): container finished" podID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerID="a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b" exitCode=0 Nov 27 12:48:32 crc kubenswrapper[4796]: I1127 12:48:32.161051 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerDied","Data":"a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b"} Nov 27 12:48:33 crc kubenswrapper[4796]: I1127 12:48:33.185033 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerStarted","Data":"3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf"} Nov 27 12:48:33 crc kubenswrapper[4796]: I1127 12:48:33.218815 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xtg9m" podStartSLOduration=3.582829242 podStartE2EDuration="7.218793529s" podCreationTimestamp="2025-11-27 12:48:26 +0000 UTC" firstStartedPulling="2025-11-27 12:48:29.131700218 +0000 UTC m=+5026.650019156" lastFinishedPulling="2025-11-27 12:48:32.767664485 +0000 UTC m=+5030.285983443" observedRunningTime="2025-11-27 12:48:33.207551522 +0000 UTC m=+5030.725870480" watchObservedRunningTime="2025-11-27 12:48:33.218793529 +0000 UTC m=+5030.737112457" Nov 27 12:48:37 crc kubenswrapper[4796]: I1127 12:48:37.369669 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:37 crc kubenswrapper[4796]: I1127 12:48:37.370283 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:37 crc kubenswrapper[4796]: I1127 12:48:37.431694 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:38 crc kubenswrapper[4796]: I1127 12:48:38.315438 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:38 crc kubenswrapper[4796]: I1127 12:48:38.378129 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xtg9m"] Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.094151 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5vdz9"] Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.099793 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.106810 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5vdz9"] Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.224876 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-catalog-content\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.224957 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jm6l\" (UniqueName: \"kubernetes.io/projected/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-kube-api-access-9jm6l\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.225297 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-utilities\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.257709 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xtg9m" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="registry-server" containerID="cri-o://3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf" gracePeriod=2 Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.327727 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-catalog-content\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.328369 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-catalog-content\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.328423 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jm6l\" (UniqueName: \"kubernetes.io/projected/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-kube-api-access-9jm6l\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.328538 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-utilities\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.329178 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-utilities\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.355026 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jm6l\" (UniqueName: \"kubernetes.io/projected/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-kube-api-access-9jm6l\") pod \"redhat-marketplace-5vdz9\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.438319 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.707815 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.841600 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-catalog-content\") pod \"eef7b83c-fe8c-4efa-8658-34555eab03c8\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.842088 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvqzs\" (UniqueName: \"kubernetes.io/projected/eef7b83c-fe8c-4efa-8658-34555eab03c8-kube-api-access-bvqzs\") pod \"eef7b83c-fe8c-4efa-8658-34555eab03c8\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.842162 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-utilities\") pod \"eef7b83c-fe8c-4efa-8658-34555eab03c8\" (UID: \"eef7b83c-fe8c-4efa-8658-34555eab03c8\") " Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.843145 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-utilities" (OuterVolumeSpecName: "utilities") pod "eef7b83c-fe8c-4efa-8658-34555eab03c8" (UID: "eef7b83c-fe8c-4efa-8658-34555eab03c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.849032 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eef7b83c-fe8c-4efa-8658-34555eab03c8-kube-api-access-bvqzs" (OuterVolumeSpecName: "kube-api-access-bvqzs") pod "eef7b83c-fe8c-4efa-8658-34555eab03c8" (UID: "eef7b83c-fe8c-4efa-8658-34555eab03c8"). InnerVolumeSpecName "kube-api-access-bvqzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.893049 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eef7b83c-fe8c-4efa-8658-34555eab03c8" (UID: "eef7b83c-fe8c-4efa-8658-34555eab03c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.944499 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvqzs\" (UniqueName: \"kubernetes.io/projected/eef7b83c-fe8c-4efa-8658-34555eab03c8-kube-api-access-bvqzs\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.944537 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.944550 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eef7b83c-fe8c-4efa-8658-34555eab03c8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:40 crc kubenswrapper[4796]: I1127 12:48:40.950789 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5vdz9"] Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.267418 4796 generic.go:334] "Generic (PLEG): container finished" podID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerID="07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc" exitCode=0 Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.267480 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5vdz9" event={"ID":"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d","Type":"ContainerDied","Data":"07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc"} Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.267504 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5vdz9" event={"ID":"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d","Type":"ContainerStarted","Data":"4d63283f0291a7282dbff280fc03b39fe82b9658982ac97316fa32a9c06a06a6"} Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.269795 4796 generic.go:334] "Generic (PLEG): container finished" podID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerID="3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf" exitCode=0 Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.269818 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerDied","Data":"3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf"} Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.269836 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtg9m" event={"ID":"eef7b83c-fe8c-4efa-8658-34555eab03c8","Type":"ContainerDied","Data":"9cee71173bb458d8f5eb887863097ffe53e74f400499b2df08d7c30978482ad0"} Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.269852 4796 scope.go:117] "RemoveContainer" containerID="3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.269965 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtg9m" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.308006 4796 scope.go:117] "RemoveContainer" containerID="a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.327817 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xtg9m"] Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.336344 4796 scope.go:117] "RemoveContainer" containerID="9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.336768 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xtg9m"] Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.353212 4796 scope.go:117] "RemoveContainer" containerID="3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf" Nov 27 12:48:41 crc kubenswrapper[4796]: E1127 12:48:41.353717 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf\": container with ID starting with 3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf not found: ID does not exist" containerID="3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.353773 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf"} err="failed to get container status \"3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf\": rpc error: code = NotFound desc = could not find container \"3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf\": container with ID starting with 3733b59bd51c17c72505bbb4f5ed8e0503afaac034f43cc352a4a89c489ddfaf not found: ID does not exist" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.353809 4796 scope.go:117] "RemoveContainer" containerID="a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b" Nov 27 12:48:41 crc kubenswrapper[4796]: E1127 12:48:41.365348 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b\": container with ID starting with a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b not found: ID does not exist" containerID="a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.365427 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b"} err="failed to get container status \"a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b\": rpc error: code = NotFound desc = could not find container \"a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b\": container with ID starting with a35c0869bf9de89479519c3a1d91191aa2acb66eabb9bc281b5ec5a2910acf4b not found: ID does not exist" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.365468 4796 scope.go:117] "RemoveContainer" containerID="9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7" Nov 27 12:48:41 crc kubenswrapper[4796]: E1127 12:48:41.366012 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7\": container with ID starting with 9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7 not found: ID does not exist" containerID="9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.366062 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7"} err="failed to get container status \"9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7\": rpc error: code = NotFound desc = could not find container \"9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7\": container with ID starting with 9bd46634ffe15f7c7681396fd18a489ed2e840c0c3fe4623741abc97ead27cc7 not found: ID does not exist" Nov 27 12:48:41 crc kubenswrapper[4796]: I1127 12:48:41.579742 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" path="/var/lib/kubelet/pods/eef7b83c-fe8c-4efa-8658-34555eab03c8/volumes" Nov 27 12:48:43 crc kubenswrapper[4796]: I1127 12:48:43.294809 4796 generic.go:334] "Generic (PLEG): container finished" podID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerID="2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77" exitCode=0 Nov 27 12:48:43 crc kubenswrapper[4796]: I1127 12:48:43.294868 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5vdz9" event={"ID":"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d","Type":"ContainerDied","Data":"2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77"} Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.309866 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5vdz9" event={"ID":"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d","Type":"ContainerStarted","Data":"a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd"} Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.354640 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5vdz9" podStartSLOduration=1.896414138 podStartE2EDuration="4.354617232s" podCreationTimestamp="2025-11-27 12:48:40 +0000 UTC" firstStartedPulling="2025-11-27 12:48:41.283570956 +0000 UTC m=+5038.801889874" lastFinishedPulling="2025-11-27 12:48:43.74177404 +0000 UTC m=+5041.260092968" observedRunningTime="2025-11-27 12:48:44.345718458 +0000 UTC m=+5041.864037376" watchObservedRunningTime="2025-11-27 12:48:44.354617232 +0000 UTC m=+5041.872936160" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.389364 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wl8wx/must-gather-trps2"] Nov 27 12:48:44 crc kubenswrapper[4796]: E1127 12:48:44.389955 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="extract-content" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.389984 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="extract-content" Nov 27 12:48:44 crc kubenswrapper[4796]: E1127 12:48:44.390034 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="extract-utilities" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.390045 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="extract-utilities" Nov 27 12:48:44 crc kubenswrapper[4796]: E1127 12:48:44.390061 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="registry-server" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.390068 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="registry-server" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.390290 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="eef7b83c-fe8c-4efa-8658-34555eab03c8" containerName="registry-server" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.391411 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.394679 4796 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wl8wx"/"default-dockercfg-zwbgl" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.394948 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wl8wx"/"openshift-service-ca.crt" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.395176 4796 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wl8wx"/"kube-root-ca.crt" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.400638 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wl8wx/must-gather-trps2"] Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.517856 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e53916a7-79e8-4b2f-886b-4278693b1877-must-gather-output\") pod \"must-gather-trps2\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.517926 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm8jk\" (UniqueName: \"kubernetes.io/projected/e53916a7-79e8-4b2f-886b-4278693b1877-kube-api-access-xm8jk\") pod \"must-gather-trps2\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.619404 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e53916a7-79e8-4b2f-886b-4278693b1877-must-gather-output\") pod \"must-gather-trps2\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.619473 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm8jk\" (UniqueName: \"kubernetes.io/projected/e53916a7-79e8-4b2f-886b-4278693b1877-kube-api-access-xm8jk\") pod \"must-gather-trps2\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.619815 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e53916a7-79e8-4b2f-886b-4278693b1877-must-gather-output\") pod \"must-gather-trps2\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.647216 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm8jk\" (UniqueName: \"kubernetes.io/projected/e53916a7-79e8-4b2f-886b-4278693b1877-kube-api-access-xm8jk\") pod \"must-gather-trps2\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:44 crc kubenswrapper[4796]: I1127 12:48:44.717603 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:48:45 crc kubenswrapper[4796]: I1127 12:48:45.373410 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wl8wx/must-gather-trps2"] Nov 27 12:48:45 crc kubenswrapper[4796]: W1127 12:48:45.776383 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode53916a7_79e8_4b2f_886b_4278693b1877.slice/crio-fbecfecb50e999b07b6ef88153b704cf79b9c4d012aa1545bbffa80f321a978f WatchSource:0}: Error finding container fbecfecb50e999b07b6ef88153b704cf79b9c4d012aa1545bbffa80f321a978f: Status 404 returned error can't find the container with id fbecfecb50e999b07b6ef88153b704cf79b9c4d012aa1545bbffa80f321a978f Nov 27 12:48:46 crc kubenswrapper[4796]: I1127 12:48:46.335920 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/must-gather-trps2" event={"ID":"e53916a7-79e8-4b2f-886b-4278693b1877","Type":"ContainerStarted","Data":"ebd9f96569470226e2bc2b4d33ab8aea643871d15db71af191863ad3bd4a923e"} Nov 27 12:48:46 crc kubenswrapper[4796]: I1127 12:48:46.336348 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/must-gather-trps2" event={"ID":"e53916a7-79e8-4b2f-886b-4278693b1877","Type":"ContainerStarted","Data":"fbecfecb50e999b07b6ef88153b704cf79b9c4d012aa1545bbffa80f321a978f"} Nov 27 12:48:47 crc kubenswrapper[4796]: I1127 12:48:47.346913 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/must-gather-trps2" event={"ID":"e53916a7-79e8-4b2f-886b-4278693b1877","Type":"ContainerStarted","Data":"9c8a6427d4e75c7a5e223aa153c05240bcf88f53005d894dd10797a892283a3f"} Nov 27 12:48:49 crc kubenswrapper[4796]: I1127 12:48:49.805301 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wl8wx/must-gather-trps2" podStartSLOduration=5.805260246 podStartE2EDuration="5.805260246s" podCreationTimestamp="2025-11-27 12:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:48:47.370209883 +0000 UTC m=+5044.888528801" watchObservedRunningTime="2025-11-27 12:48:49.805260246 +0000 UTC m=+5047.323579164" Nov 27 12:48:49 crc kubenswrapper[4796]: I1127 12:48:49.813616 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wl8wx/crc-debug-z6p25"] Nov 27 12:48:49 crc kubenswrapper[4796]: I1127 12:48:49.815212 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:49 crc kubenswrapper[4796]: I1127 12:48:49.946644 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b88ba61a-3be6-4175-9ddd-14858b18437f-host\") pod \"crc-debug-z6p25\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:49 crc kubenswrapper[4796]: I1127 12:48:49.946952 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtml2\" (UniqueName: \"kubernetes.io/projected/b88ba61a-3be6-4175-9ddd-14858b18437f-kube-api-access-mtml2\") pod \"crc-debug-z6p25\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.049138 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b88ba61a-3be6-4175-9ddd-14858b18437f-host\") pod \"crc-debug-z6p25\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.049319 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtml2\" (UniqueName: \"kubernetes.io/projected/b88ba61a-3be6-4175-9ddd-14858b18437f-kube-api-access-mtml2\") pod \"crc-debug-z6p25\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.049349 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b88ba61a-3be6-4175-9ddd-14858b18437f-host\") pod \"crc-debug-z6p25\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.081481 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtml2\" (UniqueName: \"kubernetes.io/projected/b88ba61a-3be6-4175-9ddd-14858b18437f-kube-api-access-mtml2\") pod \"crc-debug-z6p25\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.135888 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.392622 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" event={"ID":"b88ba61a-3be6-4175-9ddd-14858b18437f","Type":"ContainerStarted","Data":"e2fe63beab5fe7f7492ceb6bec49df4a72d8706ba443c2374e88c6de59d36417"} Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.438774 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.438815 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:50 crc kubenswrapper[4796]: I1127 12:48:50.490906 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:51 crc kubenswrapper[4796]: I1127 12:48:51.402439 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" event={"ID":"b88ba61a-3be6-4175-9ddd-14858b18437f","Type":"ContainerStarted","Data":"0baa95d5ca3b3ca71ebf33c9994470ac2ce386ec9de785c3ff1a377ecc251fb2"} Nov 27 12:48:51 crc kubenswrapper[4796]: I1127 12:48:51.413651 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" podStartSLOduration=2.413629904 podStartE2EDuration="2.413629904s" podCreationTimestamp="2025-11-27 12:48:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 12:48:51.41348109 +0000 UTC m=+5048.931800008" watchObservedRunningTime="2025-11-27 12:48:51.413629904 +0000 UTC m=+5048.931948822" Nov 27 12:48:51 crc kubenswrapper[4796]: I1127 12:48:51.456919 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:51 crc kubenswrapper[4796]: I1127 12:48:51.508511 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5vdz9"] Nov 27 12:48:53 crc kubenswrapper[4796]: I1127 12:48:53.418558 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5vdz9" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="registry-server" containerID="cri-o://a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd" gracePeriod=2 Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.379530 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.430604 4796 generic.go:334] "Generic (PLEG): container finished" podID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerID="a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd" exitCode=0 Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.430868 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5vdz9" event={"ID":"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d","Type":"ContainerDied","Data":"a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd"} Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.430896 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5vdz9" event={"ID":"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d","Type":"ContainerDied","Data":"4d63283f0291a7282dbff280fc03b39fe82b9658982ac97316fa32a9c06a06a6"} Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.430915 4796 scope.go:117] "RemoveContainer" containerID="a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.431086 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5vdz9" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.454348 4796 scope.go:117] "RemoveContainer" containerID="2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.503409 4796 scope.go:117] "RemoveContainer" containerID="07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.533079 4796 scope.go:117] "RemoveContainer" containerID="a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd" Nov 27 12:48:54 crc kubenswrapper[4796]: E1127 12:48:54.533621 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd\": container with ID starting with a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd not found: ID does not exist" containerID="a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.533683 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd"} err="failed to get container status \"a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd\": rpc error: code = NotFound desc = could not find container \"a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd\": container with ID starting with a45433a33ce7ff0bcbfe7d88124d83817be46054b6479c53af998af32758e1bd not found: ID does not exist" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.533716 4796 scope.go:117] "RemoveContainer" containerID="2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77" Nov 27 12:48:54 crc kubenswrapper[4796]: E1127 12:48:54.534017 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77\": container with ID starting with 2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77 not found: ID does not exist" containerID="2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.534067 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77"} err="failed to get container status \"2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77\": rpc error: code = NotFound desc = could not find container \"2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77\": container with ID starting with 2bab3e4a05cd6316833ac4df6df56292d869c7490dfd0cd2dac1e42296c70e77 not found: ID does not exist" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.534096 4796 scope.go:117] "RemoveContainer" containerID="07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc" Nov 27 12:48:54 crc kubenswrapper[4796]: E1127 12:48:54.534304 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc\": container with ID starting with 07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc not found: ID does not exist" containerID="07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.534323 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc"} err="failed to get container status \"07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc\": rpc error: code = NotFound desc = could not find container \"07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc\": container with ID starting with 07a437b79a59959b3054463efa573dbe0a9c3eca24a58330de5623ba9a9e6bbc not found: ID does not exist" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.539225 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-utilities\") pod \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.539288 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-catalog-content\") pod \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.539354 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jm6l\" (UniqueName: \"kubernetes.io/projected/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-kube-api-access-9jm6l\") pod \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\" (UID: \"51c4a0f0-0e4c-4766-a8d6-6035a9022f8d\") " Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.540254 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-utilities" (OuterVolumeSpecName: "utilities") pod "51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" (UID: "51c4a0f0-0e4c-4766-a8d6-6035a9022f8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.547773 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-kube-api-access-9jm6l" (OuterVolumeSpecName: "kube-api-access-9jm6l") pod "51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" (UID: "51c4a0f0-0e4c-4766-a8d6-6035a9022f8d"). InnerVolumeSpecName "kube-api-access-9jm6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.561626 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" (UID: "51c4a0f0-0e4c-4766-a8d6-6035a9022f8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.641834 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.642164 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.642177 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jm6l\" (UniqueName: \"kubernetes.io/projected/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d-kube-api-access-9jm6l\") on node \"crc\" DevicePath \"\"" Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.773737 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5vdz9"] Nov 27 12:48:54 crc kubenswrapper[4796]: I1127 12:48:54.784721 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5vdz9"] Nov 27 12:48:55 crc kubenswrapper[4796]: I1127 12:48:55.581352 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" path="/var/lib/kubelet/pods/51c4a0f0-0e4c-4766-a8d6-6035a9022f8d/volumes" Nov 27 12:49:01 crc kubenswrapper[4796]: I1127 12:49:01.879797 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:49:01 crc kubenswrapper[4796]: I1127 12:49:01.881094 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:49:04 crc kubenswrapper[4796]: I1127 12:49:04.575555 4796 generic.go:334] "Generic (PLEG): container finished" podID="b88ba61a-3be6-4175-9ddd-14858b18437f" containerID="0baa95d5ca3b3ca71ebf33c9994470ac2ce386ec9de785c3ff1a377ecc251fb2" exitCode=0 Nov 27 12:49:04 crc kubenswrapper[4796]: I1127 12:49:04.575610 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" event={"ID":"b88ba61a-3be6-4175-9ddd-14858b18437f","Type":"ContainerDied","Data":"0baa95d5ca3b3ca71ebf33c9994470ac2ce386ec9de785c3ff1a377ecc251fb2"} Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.272617 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.318197 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wl8wx/crc-debug-z6p25"] Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.328599 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wl8wx/crc-debug-z6p25"] Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.363756 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b88ba61a-3be6-4175-9ddd-14858b18437f-host\") pod \"b88ba61a-3be6-4175-9ddd-14858b18437f\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.363893 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b88ba61a-3be6-4175-9ddd-14858b18437f-host" (OuterVolumeSpecName: "host") pod "b88ba61a-3be6-4175-9ddd-14858b18437f" (UID: "b88ba61a-3be6-4175-9ddd-14858b18437f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.363992 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtml2\" (UniqueName: \"kubernetes.io/projected/b88ba61a-3be6-4175-9ddd-14858b18437f-kube-api-access-mtml2\") pod \"b88ba61a-3be6-4175-9ddd-14858b18437f\" (UID: \"b88ba61a-3be6-4175-9ddd-14858b18437f\") " Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.364503 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b88ba61a-3be6-4175-9ddd-14858b18437f-host\") on node \"crc\" DevicePath \"\"" Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.371753 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b88ba61a-3be6-4175-9ddd-14858b18437f-kube-api-access-mtml2" (OuterVolumeSpecName: "kube-api-access-mtml2") pod "b88ba61a-3be6-4175-9ddd-14858b18437f" (UID: "b88ba61a-3be6-4175-9ddd-14858b18437f"). InnerVolumeSpecName "kube-api-access-mtml2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.467013 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtml2\" (UniqueName: \"kubernetes.io/projected/b88ba61a-3be6-4175-9ddd-14858b18437f-kube-api-access-mtml2\") on node \"crc\" DevicePath \"\"" Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.594674 4796 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2fe63beab5fe7f7492ceb6bec49df4a72d8706ba443c2374e88c6de59d36417" Nov 27 12:49:06 crc kubenswrapper[4796]: I1127 12:49:06.594760 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-z6p25" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.501736 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wl8wx/crc-debug-94vvd"] Nov 27 12:49:07 crc kubenswrapper[4796]: E1127 12:49:07.502443 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88ba61a-3be6-4175-9ddd-14858b18437f" containerName="container-00" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.502455 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88ba61a-3be6-4175-9ddd-14858b18437f" containerName="container-00" Nov 27 12:49:07 crc kubenswrapper[4796]: E1127 12:49:07.502471 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="registry-server" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.502477 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="registry-server" Nov 27 12:49:07 crc kubenswrapper[4796]: E1127 12:49:07.502492 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="extract-content" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.502498 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="extract-content" Nov 27 12:49:07 crc kubenswrapper[4796]: E1127 12:49:07.502533 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="extract-utilities" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.502539 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="extract-utilities" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.502813 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c4a0f0-0e4c-4766-a8d6-6035a9022f8d" containerName="registry-server" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.502826 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88ba61a-3be6-4175-9ddd-14858b18437f" containerName="container-00" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.503536 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.581845 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b88ba61a-3be6-4175-9ddd-14858b18437f" path="/var/lib/kubelet/pods/b88ba61a-3be6-4175-9ddd-14858b18437f/volumes" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.588679 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c121e900-d75d-45cc-94ea-fed1ff64029c-host\") pod \"crc-debug-94vvd\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.588998 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7rb\" (UniqueName: \"kubernetes.io/projected/c121e900-d75d-45cc-94ea-fed1ff64029c-kube-api-access-8z7rb\") pod \"crc-debug-94vvd\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.691540 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c121e900-d75d-45cc-94ea-fed1ff64029c-host\") pod \"crc-debug-94vvd\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.691728 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7rb\" (UniqueName: \"kubernetes.io/projected/c121e900-d75d-45cc-94ea-fed1ff64029c-kube-api-access-8z7rb\") pod \"crc-debug-94vvd\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.691763 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c121e900-d75d-45cc-94ea-fed1ff64029c-host\") pod \"crc-debug-94vvd\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.719251 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7rb\" (UniqueName: \"kubernetes.io/projected/c121e900-d75d-45cc-94ea-fed1ff64029c-kube-api-access-8z7rb\") pod \"crc-debug-94vvd\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:07 crc kubenswrapper[4796]: I1127 12:49:07.837239 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:08 crc kubenswrapper[4796]: I1127 12:49:08.614926 4796 generic.go:334] "Generic (PLEG): container finished" podID="c121e900-d75d-45cc-94ea-fed1ff64029c" containerID="f6a1c30efa652208fba9571811d51e298661e5765218aa8c508a9763cae7b01c" exitCode=1 Nov 27 12:49:08 crc kubenswrapper[4796]: I1127 12:49:08.615017 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/crc-debug-94vvd" event={"ID":"c121e900-d75d-45cc-94ea-fed1ff64029c","Type":"ContainerDied","Data":"f6a1c30efa652208fba9571811d51e298661e5765218aa8c508a9763cae7b01c"} Nov 27 12:49:08 crc kubenswrapper[4796]: I1127 12:49:08.615503 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/crc-debug-94vvd" event={"ID":"c121e900-d75d-45cc-94ea-fed1ff64029c","Type":"ContainerStarted","Data":"78bc50a48cfd6f6e14054590a969fd1af5d713b5d460690a683a7fbd11c826a0"} Nov 27 12:49:08 crc kubenswrapper[4796]: I1127 12:49:08.659353 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wl8wx/crc-debug-94vvd"] Nov 27 12:49:08 crc kubenswrapper[4796]: I1127 12:49:08.669253 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wl8wx/crc-debug-94vvd"] Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.732606 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.833677 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c121e900-d75d-45cc-94ea-fed1ff64029c-host\") pod \"c121e900-d75d-45cc-94ea-fed1ff64029c\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.833811 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c121e900-d75d-45cc-94ea-fed1ff64029c-host" (OuterVolumeSpecName: "host") pod "c121e900-d75d-45cc-94ea-fed1ff64029c" (UID: "c121e900-d75d-45cc-94ea-fed1ff64029c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.833863 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z7rb\" (UniqueName: \"kubernetes.io/projected/c121e900-d75d-45cc-94ea-fed1ff64029c-kube-api-access-8z7rb\") pod \"c121e900-d75d-45cc-94ea-fed1ff64029c\" (UID: \"c121e900-d75d-45cc-94ea-fed1ff64029c\") " Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.834480 4796 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c121e900-d75d-45cc-94ea-fed1ff64029c-host\") on node \"crc\" DevicePath \"\"" Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.841856 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c121e900-d75d-45cc-94ea-fed1ff64029c-kube-api-access-8z7rb" (OuterVolumeSpecName: "kube-api-access-8z7rb") pod "c121e900-d75d-45cc-94ea-fed1ff64029c" (UID: "c121e900-d75d-45cc-94ea-fed1ff64029c"). InnerVolumeSpecName "kube-api-access-8z7rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:49:09 crc kubenswrapper[4796]: I1127 12:49:09.935991 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z7rb\" (UniqueName: \"kubernetes.io/projected/c121e900-d75d-45cc-94ea-fed1ff64029c-kube-api-access-8z7rb\") on node \"crc\" DevicePath \"\"" Nov 27 12:49:10 crc kubenswrapper[4796]: I1127 12:49:10.636104 4796 scope.go:117] "RemoveContainer" containerID="f6a1c30efa652208fba9571811d51e298661e5765218aa8c508a9763cae7b01c" Nov 27 12:49:10 crc kubenswrapper[4796]: I1127 12:49:10.636136 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/crc-debug-94vvd" Nov 27 12:49:11 crc kubenswrapper[4796]: I1127 12:49:11.582008 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c121e900-d75d-45cc-94ea-fed1ff64029c" path="/var/lib/kubelet/pods/c121e900-d75d-45cc-94ea-fed1ff64029c/volumes" Nov 27 12:49:31 crc kubenswrapper[4796]: I1127 12:49:31.879327 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:49:31 crc kubenswrapper[4796]: I1127 12:49:31.880054 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:50:01 crc kubenswrapper[4796]: I1127 12:50:01.879668 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:50:01 crc kubenswrapper[4796]: I1127 12:50:01.880178 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:50:01 crc kubenswrapper[4796]: I1127 12:50:01.880227 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:50:01 crc kubenswrapper[4796]: I1127 12:50:01.881045 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf308b662ff0de97573e44404d006635d7780706d2dd1c7c90a2c1aba1c97d5c"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:50:01 crc kubenswrapper[4796]: I1127 12:50:01.881100 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://cf308b662ff0de97573e44404d006635d7780706d2dd1c7c90a2c1aba1c97d5c" gracePeriod=600 Nov 27 12:50:02 crc kubenswrapper[4796]: I1127 12:50:02.142260 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="cf308b662ff0de97573e44404d006635d7780706d2dd1c7c90a2c1aba1c97d5c" exitCode=0 Nov 27 12:50:02 crc kubenswrapper[4796]: I1127 12:50:02.142779 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"cf308b662ff0de97573e44404d006635d7780706d2dd1c7c90a2c1aba1c97d5c"} Nov 27 12:50:02 crc kubenswrapper[4796]: I1127 12:50:02.142828 4796 scope.go:117] "RemoveContainer" containerID="bfb9dcf2908cd6cab06611b5e5b98277b281bf6129254275bd0adff145e35648" Nov 27 12:50:03 crc kubenswrapper[4796]: I1127 12:50:03.155337 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944"} Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.004211 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/init-config-reloader/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.185800 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/init-config-reloader/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.187566 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/config-reloader/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.193467 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_4924a8c5-467a-47c0-9763-b10559f40519/alertmanager/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.366245 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-api/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.412443 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-listener/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.418763 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-evaluator/0.log" Nov 27 12:50:13 crc kubenswrapper[4796]: I1127 12:50:13.457454 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_85e2137f-a13a-480b-9d77-e2bafab58d5d/aodh-notifier/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.272671 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-655d9778b6-79p2x_c71cc6f0-df01-4c3d-9962-29d71307f38a/barbican-api/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.309718 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-655d9778b6-79p2x_c71cc6f0-df01-4c3d-9962-29d71307f38a/barbican-api-log/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.455067 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79c6986754-nzz2p_c0327df8-c909-4f47-bff7-7519ded4ea2a/barbican-keystone-listener/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.494064 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79c6986754-nzz2p_c0327df8-c909-4f47-bff7-7519ded4ea2a/barbican-keystone-listener-log/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.542630 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7968f8455-ntchl_176ad434-447c-4540-ac74-d80d1aeab2fd/barbican-worker/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.692479 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7968f8455-ntchl_176ad434-447c-4540-ac74-d80d1aeab2fd/barbican-worker-log/0.log" Nov 27 12:50:14 crc kubenswrapper[4796]: I1127 12:50:14.836441 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2cdjx_05691e2e-3935-4eef-9d8b-f591516ebb3c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.142557 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/ceilometer-central-agent/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.292549 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/ceilometer-notification-agent/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.325092 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/proxy-httpd/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.436023 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_eeb4962a-1819-4562-baf3-1b27c476d5f1/sg-core/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.582298 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7ae7bbf2-68f9-489d-ab84-37202429a5fb/cinder-api/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.613016 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7ae7bbf2-68f9-489d-ab84-37202429a5fb/cinder-api-log/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.719846 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f8373371-1f42-4460-8724-668392be8975/cinder-scheduler/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.917650 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f8373371-1f42-4460-8724-668392be8975/probe/0.log" Nov 27 12:50:15 crc kubenswrapper[4796]: I1127 12:50:15.953011 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-krh4d_cb67680b-c997-4002-a238-67ff005b15f0/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.134343 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kfn44_c41ab70c-1f9c-4961-9e0c-eea2a846f125/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.214973 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-kdrpj_7dfd0463-f427-4277-b02d-444ef9bd8565/init/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.393378 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-kdrpj_7dfd0463-f427-4277-b02d-444ef9bd8565/init/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.416786 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-kdrpj_7dfd0463-f427-4277-b02d-444ef9bd8565/dnsmasq-dns/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.451092 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-r5mlm_72e23875-5131-453e-a9d5-dcdbea0b4cf9/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.595060 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1f85834f-7915-4c31-b55d-5c4948763a3c/glance-log/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.621827 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1f85834f-7915-4c31-b55d-5c4948763a3c/glance-httpd/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.683021 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f7681c6a-f516-4b70-848c-2a6ce2c2e087/glance-httpd/0.log" Nov 27 12:50:16 crc kubenswrapper[4796]: I1127 12:50:16.803942 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f7681c6a-f516-4b70-848c-2a6ce2c2e087/glance-log/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.119846 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5469764df4-8wt46_ecd41927-ca66-42f8-871b-2b872ba9c3c8/heat-engine/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.284087 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6896844f76-vnbq9_acf45387-6183-41dd-848c-160078149bfd/heat-api/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.309710 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t6bgb_77fd15c8-3d7d-4f3f-9d99-70ebd57ceca9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.386302 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5b7ccbdf6f-g754q_6e828f09-14e9-444f-9ef6-6a07f5f665e5/heat-cfnapi/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.707123 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bj5x7_62906f9c-c8ce-48d2-b38b-db707e82f24c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.801969 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-95c9f9f9b-zshfm_0e86b961-a118-4ef7-a58a-a1a350c48783/keystone-api/0.log" Nov 27 12:50:17 crc kubenswrapper[4796]: I1127 12:50:17.872778 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29404081-jlkpb_48aeb330-d74c-4814-9c4d-41e40436e8a3/keystone-cron/0.log" Nov 27 12:50:18 crc kubenswrapper[4796]: I1127 12:50:18.046536 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4k4z8_a779ff89-a04b-41a9-acfe-64e318bbf745/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:18 crc kubenswrapper[4796]: I1127 12:50:18.080923 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_82fecd54-d3de-4157-b04e-3187be01c0ad/kube-state-metrics/0.log" Nov 27 12:50:18 crc kubenswrapper[4796]: I1127 12:50:18.338765 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69bc59f7f-g45f2_cdd9f25b-34db-477b-a6c6-11a5732cacdc/neutron-api/0.log" Nov 27 12:50:18 crc kubenswrapper[4796]: I1127 12:50:18.360167 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69bc59f7f-g45f2_cdd9f25b-34db-477b-a6c6-11a5732cacdc/neutron-httpd/0.log" Nov 27 12:50:18 crc kubenswrapper[4796]: I1127 12:50:18.547930 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gl49z_213166c4-bd90-4e0f-8fbb-1164ebc31cb8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:18 crc kubenswrapper[4796]: I1127 12:50:18.810689 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d4574506-0fa9-41e7-ace2-0151a43916a5/nova-api-log/0.log" Nov 27 12:50:19 crc kubenswrapper[4796]: I1127 12:50:19.218142 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d4574506-0fa9-41e7-ace2-0151a43916a5/nova-api-api/0.log" Nov 27 12:50:19 crc kubenswrapper[4796]: I1127 12:50:19.246295 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a30067ee-e2bf-4d5f-bdcf-cff461743b62/nova-cell0-conductor-conductor/0.log" Nov 27 12:50:19 crc kubenswrapper[4796]: I1127 12:50:19.289805 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_97d375ad-8df3-4053-a31a-1483cbc5ce30/nova-cell1-conductor-conductor/0.log" Nov 27 12:50:19 crc kubenswrapper[4796]: I1127 12:50:19.645780 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-42r5d_bc31c0f1-6642-4ff6-8fca-16c7ca6b4941/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:19 crc kubenswrapper[4796]: I1127 12:50:19.764557 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_69f6622f-9286-49ea-b8b7-f56a61143e03/nova-cell1-novncproxy-novncproxy/0.log" Nov 27 12:50:19 crc kubenswrapper[4796]: I1127 12:50:19.826235 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_12250a93-d356-4097-b50e-57343bc8b2c9/nova-metadata-log/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.250115 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_b47781b1-daeb-4448-b7c5-55005abc5f2e/nova-scheduler-scheduler/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.264659 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bcc2f71a-d686-4c12-adc3-6a18ca6225f8/mysql-bootstrap/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.477372 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bcc2f71a-d686-4c12-adc3-6a18ca6225f8/mysql-bootstrap/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.503720 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bcc2f71a-d686-4c12-adc3-6a18ca6225f8/galera/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.675962 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8f3e487f-6783-499b-bd57-0a811f4ef5f8/mysql-bootstrap/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.883255 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8f3e487f-6783-499b-bd57-0a811f4ef5f8/mysql-bootstrap/0.log" Nov 27 12:50:20 crc kubenswrapper[4796]: I1127 12:50:20.894967 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8f3e487f-6783-499b-bd57-0a811f4ef5f8/galera/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.256387 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4bdf142a-761b-4879-b80b-b23c3f974593/openstackclient/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.356853 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-q2vvb_d5cd65c9-3282-4039-9963-f1174a310831/openstack-network-exporter/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.396906 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_12250a93-d356-4097-b50e-57343bc8b2c9/nova-metadata-metadata/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.557916 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovsdb-server-init/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.803472 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovs-vswitchd/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.860847 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovsdb-server/0.log" Nov 27 12:50:21 crc kubenswrapper[4796]: I1127 12:50:21.873510 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7cdqc_a6cbbb73-5031-4929-b184-81a8cdbc0610/ovsdb-server-init/0.log" Nov 27 12:50:22 crc kubenswrapper[4796]: I1127 12:50:22.333457 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zjmwj_99ef6c07-3c37-4a97-8fbc-bd09a91617a0/ovn-controller/0.log" Nov 27 12:50:22 crc kubenswrapper[4796]: I1127 12:50:22.444402 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7gg7l_4b5e628e-9f89-4ae6-80c0-74d9cf57535d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:22 crc kubenswrapper[4796]: I1127 12:50:22.586557 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_469ea57e-e812-49e4-a3f3-947bf53deb22/openstack-network-exporter/0.log" Nov 27 12:50:22 crc kubenswrapper[4796]: I1127 12:50:22.627767 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_469ea57e-e812-49e4-a3f3-947bf53deb22/ovn-northd/0.log" Nov 27 12:50:22 crc kubenswrapper[4796]: I1127 12:50:22.716807 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8999de5c-3782-44dc-b88d-29f80c31a55f/openstack-network-exporter/0.log" Nov 27 12:50:22 crc kubenswrapper[4796]: I1127 12:50:22.818453 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8999de5c-3782-44dc-b88d-29f80c31a55f/ovsdbserver-nb/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.032526 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_19c0f33c-82a7-4987-be98-e699c9d32051/ovsdbserver-sb/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.066509 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_19c0f33c-82a7-4987-be98-e699c9d32051/openstack-network-exporter/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.195993 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d4b54c868-49fn9_c9233c70-4c77-41e8-831b-682307ae7dd7/placement-api/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.290701 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/init-config-reloader/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.362859 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d4b54c868-49fn9_c9233c70-4c77-41e8-831b-682307ae7dd7/placement-log/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.541940 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/init-config-reloader/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.556497 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/config-reloader/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.624398 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/prometheus/0.log" Nov 27 12:50:23 crc kubenswrapper[4796]: I1127 12:50:23.631295 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8f59101a-4dfa-4b64-8f24-15041cd24d80/thanos-sidecar/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.409410 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4f2a8e-1ba5-4af3-8154-02a911ca9d14/setup-container/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.516868 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4f2a8e-1ba5-4af3-8154-02a911ca9d14/setup-container/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.540940 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4f2a8e-1ba5-4af3-8154-02a911ca9d14/rabbitmq/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.632586 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a340b9d9-d859-4b22-a55a-e955b7951806/setup-container/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.914492 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a340b9d9-d859-4b22-a55a-e955b7951806/rabbitmq/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.948368 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8bjnw_bf61d6a5-49e7-4dc9-be44-0292579185c1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:24 crc kubenswrapper[4796]: I1127 12:50:24.950480 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a340b9d9-d859-4b22-a55a-e955b7951806/setup-container/0.log" Nov 27 12:50:25 crc kubenswrapper[4796]: I1127 12:50:25.354403 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s45b6_9059a6c5-d130-4751-849d-dc05d1c5c037/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:25 crc kubenswrapper[4796]: I1127 12:50:25.466866 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7tgxb_81f86614-5247-4cb9-ad8c-688c70f862a6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:25 crc kubenswrapper[4796]: I1127 12:50:25.631032 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-ldbkm_eb6f1515-67d6-4070-ad15-fb26e89835d8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:25 crc kubenswrapper[4796]: I1127 12:50:25.707539 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ltbbg_6b7f9b64-f1d9-4468-aafd-30819c86ab8a/ssh-known-hosts-edpm-deployment/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.007589 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6568564dcc-mst5g_c65b890a-db52-4342-938c-d56538e76b3c/proxy-server/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.147556 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6568564dcc-mst5g_c65b890a-db52-4342-938c-d56538e76b3c/proxy-httpd/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.260348 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jw9v8_085a80b3-9f90-480a-8e26-fda7fb500c64/swift-ring-rebalance/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.587545 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-reaper/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.638454 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-auditor/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.784423 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-server/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.799670 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/account-replicator/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.839918 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-auditor/0.log" Nov 27 12:50:26 crc kubenswrapper[4796]: I1127 12:50:26.984926 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-replicator/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.021759 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-server/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.097038 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/container-updater/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.142248 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-auditor/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.215783 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-expirer/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.285642 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-replicator/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.335764 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-server/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.390595 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/object-updater/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.408221 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/rsync/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.522039 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_56f6dd56-8bbf-4610-b48c-898cda2e2aab/swift-recon-cron/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.660489 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kjm5k_72e3b20e-dc5b-432e-86bf-1b14e1f0e7f8/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:27 crc kubenswrapper[4796]: I1127 12:50:27.768978 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-k2s7m_8df61624-3a9b-4f6c-aa8a-2843e80fe64d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 12:50:36 crc kubenswrapper[4796]: I1127 12:50:36.193006 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7d6c8fb0-33f4-4be1-8489-4bbfa72a6b21/memcached/0.log" Nov 27 12:50:57 crc kubenswrapper[4796]: I1127 12:50:57.052422 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/util/0.log" Nov 27 12:50:57 crc kubenswrapper[4796]: I1127 12:50:57.223188 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/pull/0.log" Nov 27 12:50:57 crc kubenswrapper[4796]: I1127 12:50:57.234857 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/util/0.log" Nov 27 12:50:57 crc kubenswrapper[4796]: I1127 12:50:57.265520 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/pull/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.185301 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/util/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.412846 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/extract/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.474950 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53b0a0f005f794e648340c3a4615fa1d874b0823e21c96817a0f4b5793mckt7_94131852-2ba2-45de-a5d4-da20c127dd6d/pull/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.570480 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-z5fwq_ca4033f3-d225-4700-bf24-875462a1a404/kube-rbac-proxy/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.657343 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-z5fwq_ca4033f3-d225-4700-bf24-875462a1a404/manager/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.690188 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-lcrbs_154fc49c-3eca-4d26-836f-ed87d2d78716/kube-rbac-proxy/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.799997 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-lcrbs_154fc49c-3eca-4d26-836f-ed87d2d78716/manager/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.904958 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-78jwg_efe899c4-0891-480f-9aaa-f8b795706142/manager/0.log" Nov 27 12:50:58 crc kubenswrapper[4796]: I1127 12:50:58.954435 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-78jwg_efe899c4-0891-480f-9aaa-f8b795706142/kube-rbac-proxy/0.log" Nov 27 12:50:59 crc kubenswrapper[4796]: I1127 12:50:59.113663 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-zlbmz_0c41793c-bff2-4a6a-877c-489df83fa578/kube-rbac-proxy/0.log" Nov 27 12:50:59 crc kubenswrapper[4796]: I1127 12:50:59.237126 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-zqg7w_2b095ec7-98b1-4e3c-89ed-0bec53da6057/kube-rbac-proxy/0.log" Nov 27 12:50:59 crc kubenswrapper[4796]: I1127 12:50:59.247667 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-zlbmz_0c41793c-bff2-4a6a-877c-489df83fa578/manager/0.log" Nov 27 12:50:59 crc kubenswrapper[4796]: I1127 12:50:59.383211 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-zqg7w_2b095ec7-98b1-4e3c-89ed-0bec53da6057/manager/0.log" Nov 27 12:50:59 crc kubenswrapper[4796]: I1127 12:50:59.444881 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-wkg8j_24bc27f0-611d-4147-9819-6ec0eb012d81/manager/0.log" Nov 27 12:50:59 crc kubenswrapper[4796]: I1127 12:50:59.445119 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-wkg8j_24bc27f0-611d-4147-9819-6ec0eb012d81/kube-rbac-proxy/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.061356 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zzhd9_940eec7d-4d26-462b-958b-ec782cf28fdf/kube-rbac-proxy/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.294485 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-zzhd9_940eec7d-4d26-462b-958b-ec782cf28fdf/manager/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.309092 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-6fpn9_0b1822e4-eff9-4ae6-9d91-46bbef3d090b/kube-rbac-proxy/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.386054 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-6fpn9_0b1822e4-eff9-4ae6-9d91-46bbef3d090b/manager/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.533223 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-drlr5_5865e45d-8c50-45ea-9fb5-f8a14385de1a/manager/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.543655 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-drlr5_5865e45d-8c50-45ea-9fb5-f8a14385de1a/kube-rbac-proxy/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.645153 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-ll522_6858cdee-21cc-4595-905a-3cd4ce0da27c/kube-rbac-proxy/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.704826 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-ll522_6858cdee-21cc-4595-905a-3cd4ce0da27c/manager/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.825132 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-8wjn5_257e6c0f-c3fd-458e-839a-d28ca580fd13/manager/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.827219 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-8wjn5_257e6c0f-c3fd-458e-839a-d28ca580fd13/kube-rbac-proxy/0.log" Nov 27 12:51:00 crc kubenswrapper[4796]: I1127 12:51:00.991338 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-fsmqp_d8beead6-b3d4-4f4c-baae-4eaf00fe812a/kube-rbac-proxy/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.044926 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-fsmqp_d8beead6-b3d4-4f4c-baae-4eaf00fe812a/manager/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.105726 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6ks4r_2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1/kube-rbac-proxy/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.202323 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-6ks4r_2cf4a4fb-918b-4c51-9631-6aa3d8ddbaf1/manager/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.254003 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-8lq5v_25fd5653-6347-40b2-9d4e-531b58a7e6a3/kube-rbac-proxy/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.314547 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-8lq5v_25fd5653-6347-40b2-9d4e-531b58a7e6a3/manager/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.415042 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w_c2e58505-5b12-41ac-aa51-d857b46b0166/kube-rbac-proxy/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.455522 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bf4m6w_c2e58505-5b12-41ac-aa51-d857b46b0166/manager/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.777240 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-68f75d5585-g6s7q_b8c1db68-d88e-406c-a551-bfb6fe0df332/operator/0.log" Nov 27 12:51:01 crc kubenswrapper[4796]: I1127 12:51:01.995121 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rbwz6_037f8aa1-b5aa-4877-a71b-7ef5fa45e5d5/registry-server/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.114166 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-4xs4w_b0ba2f52-f61b-47b5-8739-3c248d6e012b/kube-rbac-proxy/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.233641 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-4xs4w_b0ba2f52-f61b-47b5-8739-3c248d6e012b/manager/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.269921 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-cjtbm_3189b882-80fe-4198-872d-b8b36f0cc766/kube-rbac-proxy/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.334561 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-cjtbm_3189b882-80fe-4198-872d-b8b36f0cc766/manager/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.464238 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-l2v4r_6cc8bd4b-fcdb-48a5-ad92-85097b2b31bb/operator/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.587592 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-grznn_b72bebe8-8cf0-4e6e-908b-a5095a7a9348/kube-rbac-proxy/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.682292 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-grznn_b72bebe8-8cf0-4e6e-908b-a5095a7a9348/manager/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.815348 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/kube-rbac-proxy/0.log" Nov 27 12:51:02 crc kubenswrapper[4796]: I1127 12:51:02.904758 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557965c5b6-9qjgg_12be3ae6-8893-4998-94bd-fd12c7c6640c/manager/0.log" Nov 27 12:51:03 crc kubenswrapper[4796]: I1127 12:51:03.028850 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f77465788-gr9vn_068c68ca-9efb-4a10-b647-9688a007ddf2/manager/0.log" Nov 27 12:51:03 crc kubenswrapper[4796]: I1127 12:51:03.071402 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-d8k66_240033a0-6f60-4c73-b2cb-8d4ddb91a66d/manager/0.log" Nov 27 12:51:03 crc kubenswrapper[4796]: I1127 12:51:03.082314 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-d8k66_240033a0-6f60-4c73-b2cb-8d4ddb91a66d/kube-rbac-proxy/0.log" Nov 27 12:51:03 crc kubenswrapper[4796]: I1127 12:51:03.123972 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-88bxr_4905ae06-8f67-4beb-93ca-bcd9e1dca63e/kube-rbac-proxy/0.log" Nov 27 12:51:03 crc kubenswrapper[4796]: I1127 12:51:03.220435 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-88bxr_4905ae06-8f67-4beb-93ca-bcd9e1dca63e/manager/0.log" Nov 27 12:51:22 crc kubenswrapper[4796]: I1127 12:51:22.717181 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8wmkg_15b7bb22-0ae1-46af-a1cf-4024996c996f/control-plane-machine-set-operator/0.log" Nov 27 12:51:22 crc kubenswrapper[4796]: I1127 12:51:22.885441 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bxf5l_96d741c5-ebc0-4c04-9192-c2661a5d0e0a/kube-rbac-proxy/0.log" Nov 27 12:51:22 crc kubenswrapper[4796]: I1127 12:51:22.914926 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bxf5l_96d741c5-ebc0-4c04-9192-c2661a5d0e0a/machine-api-operator/0.log" Nov 27 12:51:35 crc kubenswrapper[4796]: I1127 12:51:35.557153 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nx48n_22e38aa1-4d12-4b1f-bfef-12d303bc12ca/cert-manager-controller/0.log" Nov 27 12:51:35 crc kubenswrapper[4796]: I1127 12:51:35.892220 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-scllt_2a72ae09-3cbb-4ab3-91b3-0aab83b57b74/cert-manager-cainjector/0.log" Nov 27 12:51:35 crc kubenswrapper[4796]: I1127 12:51:35.923663 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qrd89_bfed3074-7dee-4f77-900a-8546651ce2e6/cert-manager-webhook/0.log" Nov 27 12:51:48 crc kubenswrapper[4796]: I1127 12:51:48.227504 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g8jfz_68dd98a6-a2d9-4a74-81c4-fab3ae549244/nmstate-console-plugin/0.log" Nov 27 12:51:48 crc kubenswrapper[4796]: I1127 12:51:48.351735 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vtgx5_e3cc1a65-8f8a-477c-96f1-3cd385bfa910/nmstate-handler/0.log" Nov 27 12:51:48 crc kubenswrapper[4796]: I1127 12:51:48.424015 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2tbf2_72f7a9b6-d5a1-4b35-8221-8c7afcaa203c/nmstate-metrics/0.log" Nov 27 12:51:48 crc kubenswrapper[4796]: I1127 12:51:48.427284 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-2tbf2_72f7a9b6-d5a1-4b35-8221-8c7afcaa203c/kube-rbac-proxy/0.log" Nov 27 12:51:48 crc kubenswrapper[4796]: I1127 12:51:48.591224 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-l6tgw_2e5faff6-d103-4aa7-ae65-77abe90548c5/nmstate-operator/0.log" Nov 27 12:51:48 crc kubenswrapper[4796]: I1127 12:51:48.638896 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-g7rhk_44d16d5c-7e22-4bf4-824e-9d58d750ac7d/nmstate-webhook/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.186174 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-f8gw9_f8304148-f799-4ea3-880c-7d7e5b744cef/kube-rbac-proxy/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.347073 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-f8gw9_f8304148-f799-4ea3-880c-7d7e5b744cef/controller/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.399901 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.607567 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.612744 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.625665 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.650497 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.853634 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.874000 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.879146 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:52:03 crc kubenswrapper[4796]: I1127 12:52:03.918787 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.127522 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/controller/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.133628 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-reloader/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.143772 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-frr-files/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.158990 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/cp-metrics/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.374510 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/kube-rbac-proxy-frr/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.389233 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/kube-rbac-proxy/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.405339 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/frr-metrics/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.627539 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/reloader/0.log" Nov 27 12:52:04 crc kubenswrapper[4796]: I1127 12:52:04.661421 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-85888_10c4179c-1abe-40a0-a5eb-ea7f70a4fde8/frr-k8s-webhook-server/0.log" Nov 27 12:52:05 crc kubenswrapper[4796]: I1127 12:52:05.084464 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-68bdcff486-mtshf_e27fc43a-060d-4f5f-9f52-4f75c9e4b651/manager/0.log" Nov 27 12:52:05 crc kubenswrapper[4796]: I1127 12:52:05.292921 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-57ffb79475-qkvck_c1a62d95-1792-4779-8f23-da2e94806d72/webhook-server/0.log" Nov 27 12:52:05 crc kubenswrapper[4796]: I1127 12:52:05.454003 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nd74_d6863326-623f-4817-ba2b-65f8b1887f5d/kube-rbac-proxy/0.log" Nov 27 12:52:05 crc kubenswrapper[4796]: I1127 12:52:05.740134 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4hg2d_eb8e31ff-f1a5-4a94-84db-ce9642a796b1/frr/0.log" Nov 27 12:52:05 crc kubenswrapper[4796]: I1127 12:52:05.967294 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nd74_d6863326-623f-4817-ba2b-65f8b1887f5d/speaker/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.141190 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/util/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.311645 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/util/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.312776 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/pull/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.363677 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/pull/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.508157 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/util/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.520636 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/pull/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.527658 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fchgnd_89a6f0f6-192f-43d8-8f81-5c0570369b9d/extract/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.662745 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/util/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.835320 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/pull/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.835368 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/pull/0.log" Nov 27 12:52:18 crc kubenswrapper[4796]: I1127 12:52:18.884080 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/util/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.030864 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/util/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.062084 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/pull/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.071459 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210nwxcp_076bc7ca-c3ef-4c76-b896-d4bdac2191e4/extract/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.219850 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/util/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.362705 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/util/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.425607 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/pull/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.463983 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/pull/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.669169 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/util/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.687998 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/pull/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.705869 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83csld2_00caecbc-f43a-453c-a3b3-a0bfa7c49f60/extract/0.log" Nov 27 12:52:19 crc kubenswrapper[4796]: I1127 12:52:19.857489 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-utilities/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.058463 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-utilities/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.067830 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-content/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.084997 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-content/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.212077 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-content/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.274393 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/extract-utilities/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.470747 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-utilities/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.632669 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-content/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.651921 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-utilities/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.712843 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-content/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.766080 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mdp44_8e43ee0f-5349-43b2-8d0c-d3f41ac9766a/registry-server/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.928064 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-content/0.log" Nov 27 12:52:20 crc kubenswrapper[4796]: I1127 12:52:20.928150 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/extract-utilities/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.174965 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-pnvmc_4a5aa833-5ce1-4847-bd66-45dcae6a2015/marketplace-operator/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.295252 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-utilities/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.417829 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f2pd5_54d4dfb4-bdf2-459f-8e07-5cd029849c42/registry-server/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.484012 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-content/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.503306 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-content/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.535391 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-utilities/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.693654 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-utilities/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.759717 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/extract-content/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.862892 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/extract-utilities/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.876199 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lz6dl_90010dd3-da45-452c-bbf0-d82893af6d97/registry-server/0.log" Nov 27 12:52:21 crc kubenswrapper[4796]: I1127 12:52:21.999932 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/extract-content/0.log" Nov 27 12:52:22 crc kubenswrapper[4796]: I1127 12:52:22.025617 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/extract-content/0.log" Nov 27 12:52:22 crc kubenswrapper[4796]: I1127 12:52:22.041254 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/extract-utilities/0.log" Nov 27 12:52:22 crc kubenswrapper[4796]: I1127 12:52:22.191070 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/extract-utilities/0.log" Nov 27 12:52:22 crc kubenswrapper[4796]: I1127 12:52:22.221634 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/extract-content/0.log" Nov 27 12:52:22 crc kubenswrapper[4796]: I1127 12:52:22.307488 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2m27c_9e381c75-af72-45c2-a671-b8c506d16fb6/registry-server/0.log" Nov 27 12:52:31 crc kubenswrapper[4796]: I1127 12:52:31.879156 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:52:31 crc kubenswrapper[4796]: I1127 12:52:31.880781 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:52:34 crc kubenswrapper[4796]: I1127 12:52:34.217259 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-q7hl9_12aebb1f-5956-4364-b2c9-80cf77ddb77a/prometheus-operator/0.log" Nov 27 12:52:34 crc kubenswrapper[4796]: I1127 12:52:34.400790 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67d796579b-bbzgw_e2638406-4d7c-404f-9906-544ce502974b/prometheus-operator-admission-webhook/0.log" Nov 27 12:52:34 crc kubenswrapper[4796]: I1127 12:52:34.437244 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67d796579b-n5622_bd9f4f30-7f8c-40d0-b559-368b90f90b4b/prometheus-operator-admission-webhook/0.log" Nov 27 12:52:34 crc kubenswrapper[4796]: I1127 12:52:34.567337 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-x26pv_2b928049-ffbb-4f1a-9d05-14f4bae23427/operator/0.log" Nov 27 12:52:34 crc kubenswrapper[4796]: I1127 12:52:34.659457 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-t8f4v_94471100-b73d-4ca3-be61-4eac947eae64/perses-operator/0.log" Nov 27 12:53:01 crc kubenswrapper[4796]: I1127 12:53:01.879667 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:53:01 crc kubenswrapper[4796]: I1127 12:53:01.880185 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:53:31 crc kubenswrapper[4796]: I1127 12:53:31.879454 4796 patch_prober.go:28] interesting pod/machine-config-daemon-v9kkr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 12:53:31 crc kubenswrapper[4796]: I1127 12:53:31.880149 4796 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 12:53:31 crc kubenswrapper[4796]: I1127 12:53:31.880213 4796 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" Nov 27 12:53:31 crc kubenswrapper[4796]: I1127 12:53:31.881164 4796 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944"} pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 12:53:31 crc kubenswrapper[4796]: I1127 12:53:31.881225 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerName="machine-config-daemon" containerID="cri-o://810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" gracePeriod=600 Nov 27 12:53:32 crc kubenswrapper[4796]: E1127 12:53:32.029777 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:53:32 crc kubenswrapper[4796]: I1127 12:53:32.239089 4796 generic.go:334] "Generic (PLEG): container finished" podID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" exitCode=0 Nov 27 12:53:32 crc kubenswrapper[4796]: I1127 12:53:32.239147 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerDied","Data":"810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944"} Nov 27 12:53:32 crc kubenswrapper[4796]: I1127 12:53:32.239505 4796 scope.go:117] "RemoveContainer" containerID="cf308b662ff0de97573e44404d006635d7780706d2dd1c7c90a2c1aba1c97d5c" Nov 27 12:53:32 crc kubenswrapper[4796]: I1127 12:53:32.240241 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:53:32 crc kubenswrapper[4796]: E1127 12:53:32.240584 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:53:45 crc kubenswrapper[4796]: I1127 12:53:45.572391 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:53:45 crc kubenswrapper[4796]: E1127 12:53:45.573317 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:53:58 crc kubenswrapper[4796]: I1127 12:53:58.569877 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:53:58 crc kubenswrapper[4796]: E1127 12:53:58.570616 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:54:03 crc kubenswrapper[4796]: I1127 12:54:03.609034 4796 generic.go:334] "Generic (PLEG): container finished" podID="e53916a7-79e8-4b2f-886b-4278693b1877" containerID="ebd9f96569470226e2bc2b4d33ab8aea643871d15db71af191863ad3bd4a923e" exitCode=0 Nov 27 12:54:03 crc kubenswrapper[4796]: I1127 12:54:03.609175 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wl8wx/must-gather-trps2" event={"ID":"e53916a7-79e8-4b2f-886b-4278693b1877","Type":"ContainerDied","Data":"ebd9f96569470226e2bc2b4d33ab8aea643871d15db71af191863ad3bd4a923e"} Nov 27 12:54:03 crc kubenswrapper[4796]: I1127 12:54:03.610257 4796 scope.go:117] "RemoveContainer" containerID="ebd9f96569470226e2bc2b4d33ab8aea643871d15db71af191863ad3bd4a923e" Nov 27 12:54:03 crc kubenswrapper[4796]: I1127 12:54:03.995459 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wl8wx_must-gather-trps2_e53916a7-79e8-4b2f-886b-4278693b1877/gather/0.log" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.620848 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-49tst"] Nov 27 12:54:07 crc kubenswrapper[4796]: E1127 12:54:07.622182 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c121e900-d75d-45cc-94ea-fed1ff64029c" containerName="container-00" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.622218 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="c121e900-d75d-45cc-94ea-fed1ff64029c" containerName="container-00" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.639544 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="c121e900-d75d-45cc-94ea-fed1ff64029c" containerName="container-00" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.641441 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.648294 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49tst"] Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.743458 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-catalog-content\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.744837 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-utilities\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.744994 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvk22\" (UniqueName: \"kubernetes.io/projected/bfc6e143-e786-4548-a126-f0e65c954240-kube-api-access-rvk22\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.846317 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-utilities\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.846408 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvk22\" (UniqueName: \"kubernetes.io/projected/bfc6e143-e786-4548-a126-f0e65c954240-kube-api-access-rvk22\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.846479 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-catalog-content\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.847022 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-utilities\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.847484 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-catalog-content\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.876439 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvk22\" (UniqueName: \"kubernetes.io/projected/bfc6e143-e786-4548-a126-f0e65c954240-kube-api-access-rvk22\") pod \"community-operators-49tst\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:07 crc kubenswrapper[4796]: I1127 12:54:07.966045 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:08 crc kubenswrapper[4796]: I1127 12:54:08.509484 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49tst"] Nov 27 12:54:08 crc kubenswrapper[4796]: W1127 12:54:08.510569 4796 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfc6e143_e786_4548_a126_f0e65c954240.slice/crio-53cba294858fd0e97508ea1eef0f808aa5cd7d9e100aed6142b33bf2ecdb9273 WatchSource:0}: Error finding container 53cba294858fd0e97508ea1eef0f808aa5cd7d9e100aed6142b33bf2ecdb9273: Status 404 returned error can't find the container with id 53cba294858fd0e97508ea1eef0f808aa5cd7d9e100aed6142b33bf2ecdb9273 Nov 27 12:54:08 crc kubenswrapper[4796]: I1127 12:54:08.669748 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerStarted","Data":"53cba294858fd0e97508ea1eef0f808aa5cd7d9e100aed6142b33bf2ecdb9273"} Nov 27 12:54:09 crc kubenswrapper[4796]: I1127 12:54:09.681529 4796 generic.go:334] "Generic (PLEG): container finished" podID="bfc6e143-e786-4548-a126-f0e65c954240" containerID="51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b" exitCode=0 Nov 27 12:54:09 crc kubenswrapper[4796]: I1127 12:54:09.681579 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerDied","Data":"51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b"} Nov 27 12:54:09 crc kubenswrapper[4796]: I1127 12:54:09.684347 4796 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 12:54:10 crc kubenswrapper[4796]: I1127 12:54:10.695569 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerStarted","Data":"4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6"} Nov 27 12:54:12 crc kubenswrapper[4796]: I1127 12:54:12.569495 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:54:12 crc kubenswrapper[4796]: E1127 12:54:12.570130 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:54:12 crc kubenswrapper[4796]: I1127 12:54:12.716911 4796 generic.go:334] "Generic (PLEG): container finished" podID="bfc6e143-e786-4548-a126-f0e65c954240" containerID="4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6" exitCode=0 Nov 27 12:54:12 crc kubenswrapper[4796]: I1127 12:54:12.716954 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerDied","Data":"4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6"} Nov 27 12:54:13 crc kubenswrapper[4796]: I1127 12:54:13.726725 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerStarted","Data":"4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514"} Nov 27 12:54:13 crc kubenswrapper[4796]: I1127 12:54:13.758363 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-49tst" podStartSLOduration=2.997580646 podStartE2EDuration="6.758340244s" podCreationTimestamp="2025-11-27 12:54:07 +0000 UTC" firstStartedPulling="2025-11-27 12:54:09.684066671 +0000 UTC m=+5367.202385589" lastFinishedPulling="2025-11-27 12:54:13.444826269 +0000 UTC m=+5370.963145187" observedRunningTime="2025-11-27 12:54:13.749341797 +0000 UTC m=+5371.267660715" watchObservedRunningTime="2025-11-27 12:54:13.758340244 +0000 UTC m=+5371.276659172" Nov 27 12:54:15 crc kubenswrapper[4796]: I1127 12:54:15.426384 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wl8wx/must-gather-trps2"] Nov 27 12:54:15 crc kubenswrapper[4796]: I1127 12:54:15.427589 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wl8wx/must-gather-trps2" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="copy" containerID="cri-o://9c8a6427d4e75c7a5e223aa153c05240bcf88f53005d894dd10797a892283a3f" gracePeriod=2 Nov 27 12:54:15 crc kubenswrapper[4796]: I1127 12:54:15.434562 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wl8wx/must-gather-trps2"] Nov 27 12:54:15 crc kubenswrapper[4796]: I1127 12:54:15.769319 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wl8wx_must-gather-trps2_e53916a7-79e8-4b2f-886b-4278693b1877/copy/0.log" Nov 27 12:54:15 crc kubenswrapper[4796]: I1127 12:54:15.769815 4796 generic.go:334] "Generic (PLEG): container finished" podID="e53916a7-79e8-4b2f-886b-4278693b1877" containerID="9c8a6427d4e75c7a5e223aa153c05240bcf88f53005d894dd10797a892283a3f" exitCode=143 Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.015917 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wl8wx_must-gather-trps2_e53916a7-79e8-4b2f-886b-4278693b1877/copy/0.log" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.016403 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.107452 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e53916a7-79e8-4b2f-886b-4278693b1877-must-gather-output\") pod \"e53916a7-79e8-4b2f-886b-4278693b1877\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.107547 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm8jk\" (UniqueName: \"kubernetes.io/projected/e53916a7-79e8-4b2f-886b-4278693b1877-kube-api-access-xm8jk\") pod \"e53916a7-79e8-4b2f-886b-4278693b1877\" (UID: \"e53916a7-79e8-4b2f-886b-4278693b1877\") " Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.115670 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e53916a7-79e8-4b2f-886b-4278693b1877-kube-api-access-xm8jk" (OuterVolumeSpecName: "kube-api-access-xm8jk") pod "e53916a7-79e8-4b2f-886b-4278693b1877" (UID: "e53916a7-79e8-4b2f-886b-4278693b1877"). InnerVolumeSpecName "kube-api-access-xm8jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.210075 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm8jk\" (UniqueName: \"kubernetes.io/projected/e53916a7-79e8-4b2f-886b-4278693b1877-kube-api-access-xm8jk\") on node \"crc\" DevicePath \"\"" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.278032 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e53916a7-79e8-4b2f-886b-4278693b1877-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e53916a7-79e8-4b2f-886b-4278693b1877" (UID: "e53916a7-79e8-4b2f-886b-4278693b1877"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.312783 4796 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e53916a7-79e8-4b2f-886b-4278693b1877-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.782444 4796 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wl8wx_must-gather-trps2_e53916a7-79e8-4b2f-886b-4278693b1877/copy/0.log" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.782956 4796 scope.go:117] "RemoveContainer" containerID="9c8a6427d4e75c7a5e223aa153c05240bcf88f53005d894dd10797a892283a3f" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.783056 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wl8wx/must-gather-trps2" Nov 27 12:54:16 crc kubenswrapper[4796]: I1127 12:54:16.803839 4796 scope.go:117] "RemoveContainer" containerID="ebd9f96569470226e2bc2b4d33ab8aea643871d15db71af191863ad3bd4a923e" Nov 27 12:54:17 crc kubenswrapper[4796]: I1127 12:54:17.581668 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" path="/var/lib/kubelet/pods/e53916a7-79e8-4b2f-886b-4278693b1877/volumes" Nov 27 12:54:17 crc kubenswrapper[4796]: I1127 12:54:17.966681 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:17 crc kubenswrapper[4796]: I1127 12:54:17.966722 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:18 crc kubenswrapper[4796]: I1127 12:54:18.037537 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:18 crc kubenswrapper[4796]: I1127 12:54:18.881660 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:18 crc kubenswrapper[4796]: I1127 12:54:18.934696 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49tst"] Nov 27 12:54:20 crc kubenswrapper[4796]: I1127 12:54:20.816842 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-49tst" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="registry-server" containerID="cri-o://4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514" gracePeriod=2 Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.256375 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.426621 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-utilities\") pod \"bfc6e143-e786-4548-a126-f0e65c954240\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.427121 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvk22\" (UniqueName: \"kubernetes.io/projected/bfc6e143-e786-4548-a126-f0e65c954240-kube-api-access-rvk22\") pod \"bfc6e143-e786-4548-a126-f0e65c954240\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.427259 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-catalog-content\") pod \"bfc6e143-e786-4548-a126-f0e65c954240\" (UID: \"bfc6e143-e786-4548-a126-f0e65c954240\") " Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.427789 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-utilities" (OuterVolumeSpecName: "utilities") pod "bfc6e143-e786-4548-a126-f0e65c954240" (UID: "bfc6e143-e786-4548-a126-f0e65c954240"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.433364 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfc6e143-e786-4548-a126-f0e65c954240-kube-api-access-rvk22" (OuterVolumeSpecName: "kube-api-access-rvk22") pod "bfc6e143-e786-4548-a126-f0e65c954240" (UID: "bfc6e143-e786-4548-a126-f0e65c954240"). InnerVolumeSpecName "kube-api-access-rvk22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.434374 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvk22\" (UniqueName: \"kubernetes.io/projected/bfc6e143-e786-4548-a126-f0e65c954240-kube-api-access-rvk22\") on node \"crc\" DevicePath \"\"" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.434471 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.828571 4796 generic.go:334] "Generic (PLEG): container finished" podID="bfc6e143-e786-4548-a126-f0e65c954240" containerID="4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514" exitCode=0 Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.828621 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerDied","Data":"4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514"} Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.828649 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tst" event={"ID":"bfc6e143-e786-4548-a126-f0e65c954240","Type":"ContainerDied","Data":"53cba294858fd0e97508ea1eef0f808aa5cd7d9e100aed6142b33bf2ecdb9273"} Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.828666 4796 scope.go:117] "RemoveContainer" containerID="4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.828690 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tst" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.863423 4796 scope.go:117] "RemoveContainer" containerID="4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.890492 4796 scope.go:117] "RemoveContainer" containerID="51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.937108 4796 scope.go:117] "RemoveContainer" containerID="4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514" Nov 27 12:54:21 crc kubenswrapper[4796]: E1127 12:54:21.937708 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514\": container with ID starting with 4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514 not found: ID does not exist" containerID="4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.937769 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514"} err="failed to get container status \"4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514\": rpc error: code = NotFound desc = could not find container \"4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514\": container with ID starting with 4cdcb23196abf10fa86b4972b05375e5c6446bfd9bbf5795ea2076def25c3514 not found: ID does not exist" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.937801 4796 scope.go:117] "RemoveContainer" containerID="4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6" Nov 27 12:54:21 crc kubenswrapper[4796]: E1127 12:54:21.938406 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6\": container with ID starting with 4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6 not found: ID does not exist" containerID="4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.938438 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6"} err="failed to get container status \"4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6\": rpc error: code = NotFound desc = could not find container \"4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6\": container with ID starting with 4fd3485534c45b35b4fe09692a340a642155bd95327796b9d7d1a27a15834eb6 not found: ID does not exist" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.938459 4796 scope.go:117] "RemoveContainer" containerID="51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b" Nov 27 12:54:21 crc kubenswrapper[4796]: E1127 12:54:21.938808 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b\": container with ID starting with 51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b not found: ID does not exist" containerID="51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b" Nov 27 12:54:21 crc kubenswrapper[4796]: I1127 12:54:21.938835 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b"} err="failed to get container status \"51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b\": rpc error: code = NotFound desc = could not find container \"51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b\": container with ID starting with 51a51bb0716620585ebfa0093e967a7c5872d87a1b921d97b4cf138aa09dea3b not found: ID does not exist" Nov 27 12:54:22 crc kubenswrapper[4796]: I1127 12:54:22.104047 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfc6e143-e786-4548-a126-f0e65c954240" (UID: "bfc6e143-e786-4548-a126-f0e65c954240"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:54:22 crc kubenswrapper[4796]: I1127 12:54:22.148889 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc6e143-e786-4548-a126-f0e65c954240-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:54:22 crc kubenswrapper[4796]: I1127 12:54:22.172848 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49tst"] Nov 27 12:54:22 crc kubenswrapper[4796]: I1127 12:54:22.189844 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-49tst"] Nov 27 12:54:23 crc kubenswrapper[4796]: I1127 12:54:23.576748 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:54:23 crc kubenswrapper[4796]: E1127 12:54:23.577761 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:54:23 crc kubenswrapper[4796]: I1127 12:54:23.582187 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfc6e143-e786-4548-a126-f0e65c954240" path="/var/lib/kubelet/pods/bfc6e143-e786-4548-a126-f0e65c954240/volumes" Nov 27 12:54:34 crc kubenswrapper[4796]: I1127 12:54:34.569280 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:54:34 crc kubenswrapper[4796]: E1127 12:54:34.570081 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:54:47 crc kubenswrapper[4796]: I1127 12:54:47.568953 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:54:47 crc kubenswrapper[4796]: E1127 12:54:47.569751 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:54:55 crc kubenswrapper[4796]: I1127 12:54:55.123673 4796 scope.go:117] "RemoveContainer" containerID="0baa95d5ca3b3ca71ebf33c9994470ac2ce386ec9de785c3ff1a377ecc251fb2" Nov 27 12:54:58 crc kubenswrapper[4796]: I1127 12:54:58.569883 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:54:58 crc kubenswrapper[4796]: E1127 12:54:58.570944 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:55:12 crc kubenswrapper[4796]: I1127 12:55:12.569241 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:55:12 crc kubenswrapper[4796]: E1127 12:55:12.570161 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:55:23 crc kubenswrapper[4796]: I1127 12:55:23.584177 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:55:23 crc kubenswrapper[4796]: E1127 12:55:23.585580 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:55:35 crc kubenswrapper[4796]: I1127 12:55:35.570933 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:55:35 crc kubenswrapper[4796]: E1127 12:55:35.571869 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:55:46 crc kubenswrapper[4796]: I1127 12:55:46.569167 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:55:46 crc kubenswrapper[4796]: E1127 12:55:46.569997 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:55:59 crc kubenswrapper[4796]: I1127 12:55:59.569795 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:55:59 crc kubenswrapper[4796]: E1127 12:55:59.570594 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:56:13 crc kubenswrapper[4796]: I1127 12:56:13.584101 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:56:13 crc kubenswrapper[4796]: E1127 12:56:13.585222 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:56:28 crc kubenswrapper[4796]: I1127 12:56:28.569930 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:56:28 crc kubenswrapper[4796]: E1127 12:56:28.570757 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:56:41 crc kubenswrapper[4796]: I1127 12:56:41.569468 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:56:41 crc kubenswrapper[4796]: E1127 12:56:41.570294 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:56:54 crc kubenswrapper[4796]: I1127 12:56:54.569423 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:56:54 crc kubenswrapper[4796]: E1127 12:56:54.570216 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:57:08 crc kubenswrapper[4796]: I1127 12:57:08.570075 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:57:08 crc kubenswrapper[4796]: E1127 12:57:08.571194 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:57:19 crc kubenswrapper[4796]: I1127 12:57:19.569093 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:57:19 crc kubenswrapper[4796]: E1127 12:57:19.569731 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:57:34 crc kubenswrapper[4796]: I1127 12:57:34.569744 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:57:34 crc kubenswrapper[4796]: E1127 12:57:34.570527 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:57:46 crc kubenswrapper[4796]: I1127 12:57:46.570864 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:57:46 crc kubenswrapper[4796]: E1127 12:57:46.571782 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.371393 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x8zj8"] Nov 27 12:57:52 crc kubenswrapper[4796]: E1127 12:57:52.373997 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="registry-server" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374033 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="registry-server" Nov 27 12:57:52 crc kubenswrapper[4796]: E1127 12:57:52.374068 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="extract-content" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374086 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="extract-content" Nov 27 12:57:52 crc kubenswrapper[4796]: E1127 12:57:52.374126 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="extract-utilities" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374142 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="extract-utilities" Nov 27 12:57:52 crc kubenswrapper[4796]: E1127 12:57:52.374174 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="copy" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374190 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="copy" Nov 27 12:57:52 crc kubenswrapper[4796]: E1127 12:57:52.374226 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="gather" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374242 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="gather" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374742 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="gather" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374794 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc6e143-e786-4548-a126-f0e65c954240" containerName="registry-server" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.374828 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53916a7-79e8-4b2f-886b-4278693b1877" containerName="copy" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.378187 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.401666 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8zj8"] Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.545204 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-catalog-content\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.545489 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb8vr\" (UniqueName: \"kubernetes.io/projected/ea00950f-a978-43da-aa41-d6959ac17f10-kube-api-access-pb8vr\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.546090 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-utilities\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.649351 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-utilities\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.649413 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-utilities\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.649671 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-catalog-content\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.649809 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb8vr\" (UniqueName: \"kubernetes.io/projected/ea00950f-a978-43da-aa41-d6959ac17f10-kube-api-access-pb8vr\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.650611 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-catalog-content\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.675336 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb8vr\" (UniqueName: \"kubernetes.io/projected/ea00950f-a978-43da-aa41-d6959ac17f10-kube-api-access-pb8vr\") pod \"redhat-operators-x8zj8\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:52 crc kubenswrapper[4796]: I1127 12:57:52.717897 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:57:53 crc kubenswrapper[4796]: I1127 12:57:53.185452 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8zj8"] Nov 27 12:57:53 crc kubenswrapper[4796]: I1127 12:57:53.449121 4796 generic.go:334] "Generic (PLEG): container finished" podID="ea00950f-a978-43da-aa41-d6959ac17f10" containerID="30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad" exitCode=0 Nov 27 12:57:53 crc kubenswrapper[4796]: I1127 12:57:53.449223 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerDied","Data":"30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad"} Nov 27 12:57:53 crc kubenswrapper[4796]: I1127 12:57:53.449411 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerStarted","Data":"4056d5ec3f6e184370d163fc1d8b94b2411b946332ba91d652c19308296bcbe7"} Nov 27 12:57:54 crc kubenswrapper[4796]: I1127 12:57:54.474538 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerStarted","Data":"abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9"} Nov 27 12:57:57 crc kubenswrapper[4796]: I1127 12:57:57.514066 4796 generic.go:334] "Generic (PLEG): container finished" podID="ea00950f-a978-43da-aa41-d6959ac17f10" containerID="abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9" exitCode=0 Nov 27 12:57:57 crc kubenswrapper[4796]: I1127 12:57:57.514303 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerDied","Data":"abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9"} Nov 27 12:57:58 crc kubenswrapper[4796]: I1127 12:57:58.525857 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerStarted","Data":"db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582"} Nov 27 12:57:58 crc kubenswrapper[4796]: I1127 12:57:58.551593 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x8zj8" podStartSLOduration=2.008192735 podStartE2EDuration="6.551572862s" podCreationTimestamp="2025-11-27 12:57:52 +0000 UTC" firstStartedPulling="2025-11-27 12:57:53.451249001 +0000 UTC m=+5590.969567919" lastFinishedPulling="2025-11-27 12:57:57.994629128 +0000 UTC m=+5595.512948046" observedRunningTime="2025-11-27 12:57:58.547002458 +0000 UTC m=+5596.065321386" watchObservedRunningTime="2025-11-27 12:57:58.551572862 +0000 UTC m=+5596.069891780" Nov 27 12:58:01 crc kubenswrapper[4796]: I1127 12:58:01.571198 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:58:01 crc kubenswrapper[4796]: E1127 12:58:01.571941 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:58:02 crc kubenswrapper[4796]: I1127 12:58:02.718260 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:58:02 crc kubenswrapper[4796]: I1127 12:58:02.718331 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:58:03 crc kubenswrapper[4796]: I1127 12:58:03.808194 4796 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x8zj8" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="registry-server" probeResult="failure" output=< Nov 27 12:58:03 crc kubenswrapper[4796]: timeout: failed to connect service ":50051" within 1s Nov 27 12:58:03 crc kubenswrapper[4796]: > Nov 27 12:58:12 crc kubenswrapper[4796]: I1127 12:58:12.570731 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:58:12 crc kubenswrapper[4796]: E1127 12:58:12.571684 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:58:12 crc kubenswrapper[4796]: I1127 12:58:12.803990 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:58:12 crc kubenswrapper[4796]: I1127 12:58:12.857606 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:58:13 crc kubenswrapper[4796]: I1127 12:58:13.058504 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8zj8"] Nov 27 12:58:14 crc kubenswrapper[4796]: I1127 12:58:14.719199 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x8zj8" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="registry-server" containerID="cri-o://db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582" gracePeriod=2 Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.220392 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.352120 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-utilities\") pod \"ea00950f-a978-43da-aa41-d6959ac17f10\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.352403 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-catalog-content\") pod \"ea00950f-a978-43da-aa41-d6959ac17f10\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.352618 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb8vr\" (UniqueName: \"kubernetes.io/projected/ea00950f-a978-43da-aa41-d6959ac17f10-kube-api-access-pb8vr\") pod \"ea00950f-a978-43da-aa41-d6959ac17f10\" (UID: \"ea00950f-a978-43da-aa41-d6959ac17f10\") " Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.353506 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-utilities" (OuterVolumeSpecName: "utilities") pod "ea00950f-a978-43da-aa41-d6959ac17f10" (UID: "ea00950f-a978-43da-aa41-d6959ac17f10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.359725 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea00950f-a978-43da-aa41-d6959ac17f10-kube-api-access-pb8vr" (OuterVolumeSpecName: "kube-api-access-pb8vr") pod "ea00950f-a978-43da-aa41-d6959ac17f10" (UID: "ea00950f-a978-43da-aa41-d6959ac17f10"). InnerVolumeSpecName "kube-api-access-pb8vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.455252 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb8vr\" (UniqueName: \"kubernetes.io/projected/ea00950f-a978-43da-aa41-d6959ac17f10-kube-api-access-pb8vr\") on node \"crc\" DevicePath \"\"" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.455297 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.489737 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea00950f-a978-43da-aa41-d6959ac17f10" (UID: "ea00950f-a978-43da-aa41-d6959ac17f10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.557013 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea00950f-a978-43da-aa41-d6959ac17f10-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.735208 4796 generic.go:334] "Generic (PLEG): container finished" podID="ea00950f-a978-43da-aa41-d6959ac17f10" containerID="db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582" exitCode=0 Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.735311 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8zj8" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.735319 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerDied","Data":"db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582"} Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.736475 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8zj8" event={"ID":"ea00950f-a978-43da-aa41-d6959ac17f10","Type":"ContainerDied","Data":"4056d5ec3f6e184370d163fc1d8b94b2411b946332ba91d652c19308296bcbe7"} Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.736506 4796 scope.go:117] "RemoveContainer" containerID="db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.766968 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8zj8"] Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.767833 4796 scope.go:117] "RemoveContainer" containerID="abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.780103 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x8zj8"] Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.797178 4796 scope.go:117] "RemoveContainer" containerID="30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.856354 4796 scope.go:117] "RemoveContainer" containerID="db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582" Nov 27 12:58:15 crc kubenswrapper[4796]: E1127 12:58:15.856768 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582\": container with ID starting with db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582 not found: ID does not exist" containerID="db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.856803 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582"} err="failed to get container status \"db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582\": rpc error: code = NotFound desc = could not find container \"db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582\": container with ID starting with db6e48a40f60027888c8fe4c901e712d1e51aaa4ea586b284e63011b3268c582 not found: ID does not exist" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.856827 4796 scope.go:117] "RemoveContainer" containerID="abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9" Nov 27 12:58:15 crc kubenswrapper[4796]: E1127 12:58:15.857080 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9\": container with ID starting with abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9 not found: ID does not exist" containerID="abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.857126 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9"} err="failed to get container status \"abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9\": rpc error: code = NotFound desc = could not find container \"abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9\": container with ID starting with abca2354f445c38b9fb58b39937756610c10dbf3566e90cc09c5c9a84d3a0ed9 not found: ID does not exist" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.857151 4796 scope.go:117] "RemoveContainer" containerID="30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad" Nov 27 12:58:15 crc kubenswrapper[4796]: E1127 12:58:15.857530 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad\": container with ID starting with 30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad not found: ID does not exist" containerID="30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad" Nov 27 12:58:15 crc kubenswrapper[4796]: I1127 12:58:15.857552 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad"} err="failed to get container status \"30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad\": rpc error: code = NotFound desc = could not find container \"30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad\": container with ID starting with 30dea94e28ae3e543a77e053a26db933d1dc7b4302c9464a3c2f040b39f72dad not found: ID does not exist" Nov 27 12:58:17 crc kubenswrapper[4796]: I1127 12:58:17.598724 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" path="/var/lib/kubelet/pods/ea00950f-a978-43da-aa41-d6959ac17f10/volumes" Nov 27 12:58:27 crc kubenswrapper[4796]: I1127 12:58:27.569446 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:58:27 crc kubenswrapper[4796]: E1127 12:58:27.570153 4796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v9kkr_openshift-machine-config-operator(e606fa06-e313-4bb9-b2cc-84ff65829b3c)\"" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" podUID="e606fa06-e313-4bb9-b2cc-84ff65829b3c" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.197166 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jtw8t"] Nov 27 12:58:28 crc kubenswrapper[4796]: E1127 12:58:28.198472 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="extract-content" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.198511 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="extract-content" Nov 27 12:58:28 crc kubenswrapper[4796]: E1127 12:58:28.198601 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="extract-utilities" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.198620 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="extract-utilities" Nov 27 12:58:28 crc kubenswrapper[4796]: E1127 12:58:28.198669 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="registry-server" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.198687 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="registry-server" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.199403 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea00950f-a978-43da-aa41-d6959ac17f10" containerName="registry-server" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.203799 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.229160 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtw8t"] Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.332836 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqxvq\" (UniqueName: \"kubernetes.io/projected/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-kube-api-access-xqxvq\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.332915 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-catalog-content\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.333048 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-utilities\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.434411 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqxvq\" (UniqueName: \"kubernetes.io/projected/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-kube-api-access-xqxvq\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.434480 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-catalog-content\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.434553 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-utilities\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.434992 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-utilities\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.436000 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-catalog-content\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.476430 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqxvq\" (UniqueName: \"kubernetes.io/projected/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-kube-api-access-xqxvq\") pod \"certified-operators-jtw8t\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:28 crc kubenswrapper[4796]: I1127 12:58:28.535572 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:29 crc kubenswrapper[4796]: I1127 12:58:29.060812 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtw8t"] Nov 27 12:58:29 crc kubenswrapper[4796]: I1127 12:58:29.884150 4796 generic.go:334] "Generic (PLEG): container finished" podID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerID="c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692" exitCode=0 Nov 27 12:58:29 crc kubenswrapper[4796]: I1127 12:58:29.884277 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerDied","Data":"c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692"} Nov 27 12:58:29 crc kubenswrapper[4796]: I1127 12:58:29.884842 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerStarted","Data":"4d25f2d2df095d570bf0100e48615d8b2453e7534ee587b65e1292522f3417ed"} Nov 27 12:58:31 crc kubenswrapper[4796]: I1127 12:58:31.907761 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerStarted","Data":"82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a"} Nov 27 12:58:32 crc kubenswrapper[4796]: I1127 12:58:32.917847 4796 generic.go:334] "Generic (PLEG): container finished" podID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerID="82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a" exitCode=0 Nov 27 12:58:32 crc kubenswrapper[4796]: I1127 12:58:32.917910 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerDied","Data":"82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a"} Nov 27 12:58:33 crc kubenswrapper[4796]: I1127 12:58:33.932181 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerStarted","Data":"db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c"} Nov 27 12:58:33 crc kubenswrapper[4796]: I1127 12:58:33.951770 4796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jtw8t" podStartSLOduration=2.443622131 podStartE2EDuration="5.951742569s" podCreationTimestamp="2025-11-27 12:58:28 +0000 UTC" firstStartedPulling="2025-11-27 12:58:29.885723386 +0000 UTC m=+5627.404042304" lastFinishedPulling="2025-11-27 12:58:33.393843824 +0000 UTC m=+5630.912162742" observedRunningTime="2025-11-27 12:58:33.951353768 +0000 UTC m=+5631.469672686" watchObservedRunningTime="2025-11-27 12:58:33.951742569 +0000 UTC m=+5631.470061537" Nov 27 12:58:38 crc kubenswrapper[4796]: I1127 12:58:38.537160 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:38 crc kubenswrapper[4796]: I1127 12:58:38.537811 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:38 crc kubenswrapper[4796]: I1127 12:58:38.612650 4796 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:39 crc kubenswrapper[4796]: I1127 12:58:39.085061 4796 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:39 crc kubenswrapper[4796]: I1127 12:58:39.158010 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtw8t"] Nov 27 12:58:39 crc kubenswrapper[4796]: I1127 12:58:39.569729 4796 scope.go:117] "RemoveContainer" containerID="810b52ea0006dc93693272d29319c09060dc2aada5ace56199c2d96e3506f944" Nov 27 12:58:40 crc kubenswrapper[4796]: I1127 12:58:40.017794 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v9kkr" event={"ID":"e606fa06-e313-4bb9-b2cc-84ff65829b3c","Type":"ContainerStarted","Data":"c37d4142a7ca7a92ed92ea5e9f756352af81c8069c311ef3a6e5a208cbc29f19"} Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.028974 4796 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jtw8t" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="registry-server" containerID="cri-o://db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c" gracePeriod=2 Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.522969 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.711878 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqxvq\" (UniqueName: \"kubernetes.io/projected/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-kube-api-access-xqxvq\") pod \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.712017 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-catalog-content\") pod \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.712162 4796 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-utilities\") pod \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\" (UID: \"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17\") " Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.713468 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-utilities" (OuterVolumeSpecName: "utilities") pod "6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" (UID: "6e0ef08c-b033-4297-9e6d-bdff1fd9cc17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.717993 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-kube-api-access-xqxvq" (OuterVolumeSpecName: "kube-api-access-xqxvq") pod "6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" (UID: "6e0ef08c-b033-4297-9e6d-bdff1fd9cc17"). InnerVolumeSpecName "kube-api-access-xqxvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.764222 4796 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" (UID: "6e0ef08c-b033-4297-9e6d-bdff1fd9cc17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.815075 4796 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqxvq\" (UniqueName: \"kubernetes.io/projected/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-kube-api-access-xqxvq\") on node \"crc\" DevicePath \"\"" Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.815115 4796 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 12:58:41 crc kubenswrapper[4796]: I1127 12:58:41.815128 4796 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.045987 4796 generic.go:334] "Generic (PLEG): container finished" podID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerID="db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c" exitCode=0 Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.046069 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerDied","Data":"db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c"} Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.046152 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtw8t" event={"ID":"6e0ef08c-b033-4297-9e6d-bdff1fd9cc17","Type":"ContainerDied","Data":"4d25f2d2df095d570bf0100e48615d8b2453e7534ee587b65e1292522f3417ed"} Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.046191 4796 scope.go:117] "RemoveContainer" containerID="db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.047632 4796 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtw8t" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.093049 4796 scope.go:117] "RemoveContainer" containerID="82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.094626 4796 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtw8t"] Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.104424 4796 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jtw8t"] Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.128420 4796 scope.go:117] "RemoveContainer" containerID="c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.171600 4796 scope.go:117] "RemoveContainer" containerID="db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c" Nov 27 12:58:42 crc kubenswrapper[4796]: E1127 12:58:42.172142 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c\": container with ID starting with db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c not found: ID does not exist" containerID="db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.172190 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c"} err="failed to get container status \"db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c\": rpc error: code = NotFound desc = could not find container \"db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c\": container with ID starting with db21e401b65fb289a6994c416eca0b0072092bafa947faa95a3b0eebf494c46c not found: ID does not exist" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.172222 4796 scope.go:117] "RemoveContainer" containerID="82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a" Nov 27 12:58:42 crc kubenswrapper[4796]: E1127 12:58:42.174699 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a\": container with ID starting with 82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a not found: ID does not exist" containerID="82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.174740 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a"} err="failed to get container status \"82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a\": rpc error: code = NotFound desc = could not find container \"82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a\": container with ID starting with 82155f16e217cfa1e719ea450094ec76a1a8803fe1ad00e232f576d8a2d26d4a not found: ID does not exist" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.174763 4796 scope.go:117] "RemoveContainer" containerID="c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692" Nov 27 12:58:42 crc kubenswrapper[4796]: E1127 12:58:42.175550 4796 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692\": container with ID starting with c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692 not found: ID does not exist" containerID="c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692" Nov 27 12:58:42 crc kubenswrapper[4796]: I1127 12:58:42.175601 4796 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692"} err="failed to get container status \"c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692\": rpc error: code = NotFound desc = could not find container \"c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692\": container with ID starting with c7043ea4ac2d68debee6bb8ab2c7df0d1c9af24f94febe454569d3a922c33692 not found: ID does not exist" Nov 27 12:58:43 crc kubenswrapper[4796]: I1127 12:58:43.589448 4796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" path="/var/lib/kubelet/pods/6e0ef08c-b033-4297-9e6d-bdff1fd9cc17/volumes" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.874126 4796 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4rr"] Nov 27 12:58:46 crc kubenswrapper[4796]: E1127 12:58:46.875402 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="registry-server" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.875425 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="registry-server" Nov 27 12:58:46 crc kubenswrapper[4796]: E1127 12:58:46.875451 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="extract-utilities" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.875462 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="extract-utilities" Nov 27 12:58:46 crc kubenswrapper[4796]: E1127 12:58:46.875497 4796 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="extract-content" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.875510 4796 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="extract-content" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.875786 4796 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0ef08c-b033-4297-9e6d-bdff1fd9cc17" containerName="registry-server" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.877720 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:46 crc kubenswrapper[4796]: I1127 12:58:46.887870 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4rr"] Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.025236 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6rjt\" (UniqueName: \"kubernetes.io/projected/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-kube-api-access-w6rjt\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.025648 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-utilities\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.025730 4796 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-catalog-content\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.129675 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6rjt\" (UniqueName: \"kubernetes.io/projected/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-kube-api-access-w6rjt\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.130033 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-utilities\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.130151 4796 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-catalog-content\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.131050 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-utilities\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.131159 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-catalog-content\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.161740 4796 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6rjt\" (UniqueName: \"kubernetes.io/projected/ff725a30-e2df-4b81-bbd8-00a3acc55ddf-kube-api-access-w6rjt\") pod \"redhat-marketplace-gp4rr\" (UID: \"ff725a30-e2df-4b81-bbd8-00a3acc55ddf\") " pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.208369 4796 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp4rr" Nov 27 12:58:47 crc kubenswrapper[4796]: I1127 12:58:47.654870 4796 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4rr"] Nov 27 12:58:48 crc kubenswrapper[4796]: I1127 12:58:48.151160 4796 generic.go:334] "Generic (PLEG): container finished" podID="ff725a30-e2df-4b81-bbd8-00a3acc55ddf" containerID="6cb7f3de549351dde05ff47c8c3aecf8933abb7a625a36025039b73706aab3cc" exitCode=0 Nov 27 12:58:48 crc kubenswrapper[4796]: I1127 12:58:48.151328 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4rr" event={"ID":"ff725a30-e2df-4b81-bbd8-00a3acc55ddf","Type":"ContainerDied","Data":"6cb7f3de549351dde05ff47c8c3aecf8933abb7a625a36025039b73706aab3cc"} Nov 27 12:58:48 crc kubenswrapper[4796]: I1127 12:58:48.151516 4796 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4rr" event={"ID":"ff725a30-e2df-4b81-bbd8-00a3acc55ddf","Type":"ContainerStarted","Data":"c9480e4ca2253e07df495021f79a169d1ba7bf1c2db403dec88c74ddc83d3364"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112045416024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112045416017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112032034016474 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112032035015445 5ustar corecore